mirror of https://bitbucket.org/ausocean/av.git
Merged in separate-codecs (pull request #188)
codec: created h264 and mjpeg packages to house lexers Approved-by: Alan Noble <anoble@gmail.com>
This commit is contained in:
commit
eaba622909
|
@ -3,7 +3,7 @@ NAME
|
||||||
lex.go
|
lex.go
|
||||||
|
|
||||||
DESCRIPTION
|
DESCRIPTION
|
||||||
See Readme.md
|
lex.go provides a lexer to lex h264 bytestream into access units.
|
||||||
|
|
||||||
AUTHOR
|
AUTHOR
|
||||||
Dan Kortschak <dan@ausocean.org>
|
Dan Kortschak <dan@ausocean.org>
|
||||||
|
@ -25,13 +25,11 @@ LICENSE
|
||||||
along with revid in gpl.txt. If not, see http://www.gnu.org/licenses.
|
along with revid in gpl.txt. If not, see http://www.gnu.org/licenses.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
// Package lex provides lexers for video encodings.
|
// lex.go provides a lexer to lex h264 bytestream into access units.
|
||||||
package lex
|
|
||||||
|
package h264
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bufio"
|
|
||||||
"bytes"
|
|
||||||
"fmt"
|
|
||||||
"io"
|
"io"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
@ -48,7 +46,7 @@ var h264Prefix = [...]byte{0x00, 0x00, 0x01, 0x09, 0xf0}
|
||||||
// successive writes being performed not earlier than the specified delay.
|
// successive writes being performed not earlier than the specified delay.
|
||||||
// NAL units are split after type 1 (Coded slice of a non-IDR picture), 5
|
// NAL units are split after type 1 (Coded slice of a non-IDR picture), 5
|
||||||
// (Coded slice of a IDR picture) and 8 (Picture parameter set).
|
// (Coded slice of a IDR picture) and 8 (Picture parameter set).
|
||||||
func H264(dst io.Writer, src io.Reader, delay time.Duration) error {
|
func Lex(dst io.Writer, src io.Reader, delay time.Duration) error {
|
||||||
var tick <-chan time.Time
|
var tick <-chan time.Time
|
||||||
if delay == 0 {
|
if delay == 0 {
|
||||||
tick = noDelay
|
tick = noDelay
|
||||||
|
@ -200,48 +198,3 @@ func (c *scanner) reload() error {
|
||||||
c.off = 0
|
c.off = 0
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// MJPEG parses MJPEG frames read from src into separate writes to dst with
|
|
||||||
// successive writes being performed not earlier than the specified delay.
|
|
||||||
func MJPEG(dst io.Writer, src io.Reader, delay time.Duration) error {
|
|
||||||
var tick <-chan time.Time
|
|
||||||
if delay == 0 {
|
|
||||||
tick = noDelay
|
|
||||||
} else {
|
|
||||||
ticker := time.NewTicker(delay)
|
|
||||||
defer ticker.Stop()
|
|
||||||
tick = ticker.C
|
|
||||||
}
|
|
||||||
|
|
||||||
r := bufio.NewReader(src)
|
|
||||||
for {
|
|
||||||
buf := make([]byte, 2, 4<<10)
|
|
||||||
n, err := r.Read(buf)
|
|
||||||
if n < 2 {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if !bytes.Equal(buf, []byte{0xff, 0xd8}) {
|
|
||||||
return fmt.Errorf("parser: not MJPEG frame start: %#v", buf)
|
|
||||||
}
|
|
||||||
var last byte
|
|
||||||
for {
|
|
||||||
b, err := r.ReadByte()
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
buf = append(buf, b)
|
|
||||||
if last == 0xff && b == 0xd9 {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
last = b
|
|
||||||
}
|
|
||||||
<-tick
|
|
||||||
_, err = dst.Write(buf)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -3,7 +3,7 @@ NAME
|
||||||
lex_test.go
|
lex_test.go
|
||||||
|
|
||||||
DESCRIPTION
|
DESCRIPTION
|
||||||
See Readme.md
|
lex_test.go provides tests for the lexer in lex.go.
|
||||||
|
|
||||||
AUTHOR
|
AUTHOR
|
||||||
Dan Kortschak <dan@ausocean.org>
|
Dan Kortschak <dan@ausocean.org>
|
||||||
|
@ -25,7 +25,9 @@ LICENSE
|
||||||
along with revid in gpl.txt. If not, see http://www.gnu.org/licenses.
|
along with revid in gpl.txt. If not, see http://www.gnu.org/licenses.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
package lex
|
// lex_test.go provides tests for the lexer in lex.go.
|
||||||
|
|
||||||
|
package h264
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
|
@ -207,7 +209,7 @@ var h264Tests = []struct {
|
||||||
func TestH264(t *testing.T) {
|
func TestH264(t *testing.T) {
|
||||||
for _, test := range h264Tests {
|
for _, test := range h264Tests {
|
||||||
var buf chunkEncoder
|
var buf chunkEncoder
|
||||||
err := H264(&buf, bytes.NewReader(test.input), test.delay)
|
err := Lex(&buf, bytes.NewReader(test.input), test.delay)
|
||||||
if fmt.Sprint(err) != fmt.Sprint(test.err) {
|
if fmt.Sprint(err) != fmt.Sprint(test.err) {
|
||||||
t.Errorf("unexpected error for %q: got:%v want:%v", test.name, err, test.err)
|
t.Errorf("unexpected error for %q: got:%v want:%v", test.name, err, test.err)
|
||||||
}
|
}
|
||||||
|
@ -221,87 +223,6 @@ func TestH264(t *testing.T) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
*/
|
*/
|
||||||
|
|
||||||
var mjpegTests = []struct {
|
|
||||||
name string
|
|
||||||
input []byte
|
|
||||||
delay time.Duration
|
|
||||||
want [][]byte
|
|
||||||
err error
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "empty",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "null",
|
|
||||||
input: []byte{0xff, 0xd8, 0xff, 0xd9},
|
|
||||||
delay: 0,
|
|
||||||
want: [][]byte{{0xff, 0xd8, 0xff, 0xd9}},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "null delayed",
|
|
||||||
input: []byte{0xff, 0xd8, 0xff, 0xd9},
|
|
||||||
delay: time.Millisecond,
|
|
||||||
want: [][]byte{{0xff, 0xd8, 0xff, 0xd9}},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "full",
|
|
||||||
input: []byte{
|
|
||||||
0xff, 0xd8, 'f', 'u', 'l', 'l', 0xff, 0xd9,
|
|
||||||
0xff, 0xd8, 'f', 'r', 'a', 'm', 'e', 0xff, 0xd9,
|
|
||||||
0xff, 0xd8, 'w', 'i', 't', 'h', 0xff, 0xd9,
|
|
||||||
0xff, 0xd8, 'l', 'e', 'n', 'g', 't', 'h', 0xff, 0xd9,
|
|
||||||
0xff, 0xd8, 's', 'p', 'r', 'e', 'a', 'd', 0xff, 0xd9,
|
|
||||||
},
|
|
||||||
delay: 0,
|
|
||||||
want: [][]byte{
|
|
||||||
{0xff, 0xd8, 'f', 'u', 'l', 'l', 0xff, 0xd9},
|
|
||||||
{0xff, 0xd8, 'f', 'r', 'a', 'm', 'e', 0xff, 0xd9},
|
|
||||||
{0xff, 0xd8, 'w', 'i', 't', 'h', 0xff, 0xd9},
|
|
||||||
{0xff, 0xd8, 'l', 'e', 'n', 'g', 't', 'h', 0xff, 0xd9},
|
|
||||||
{0xff, 0xd8, 's', 'p', 'r', 'e', 'a', 'd', 0xff, 0xd9},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "full delayed",
|
|
||||||
input: []byte{
|
|
||||||
0xff, 0xd8, 'f', 'u', 'l', 'l', 0xff, 0xd9,
|
|
||||||
0xff, 0xd8, 'f', 'r', 'a', 'm', 'e', 0xff, 0xd9,
|
|
||||||
0xff, 0xd8, 'w', 'i', 't', 'h', 0xff, 0xd9,
|
|
||||||
0xff, 0xd8, 'l', 'e', 'n', 'g', 't', 'h', 0xff, 0xd9,
|
|
||||||
0xff, 0xd8, 's', 'p', 'r', 'e', 'a', 'd', 0xff, 0xd9,
|
|
||||||
},
|
|
||||||
delay: time.Millisecond,
|
|
||||||
want: [][]byte{
|
|
||||||
{0xff, 0xd8, 'f', 'u', 'l', 'l', 0xff, 0xd9},
|
|
||||||
{0xff, 0xd8, 'f', 'r', 'a', 'm', 'e', 0xff, 0xd9},
|
|
||||||
{0xff, 0xd8, 'w', 'i', 't', 'h', 0xff, 0xd9},
|
|
||||||
{0xff, 0xd8, 'l', 'e', 'n', 'g', 't', 'h', 0xff, 0xd9},
|
|
||||||
{0xff, 0xd8, 's', 'p', 'r', 'e', 'a', 'd', 0xff, 0xd9},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
// FIXME this needs to be adapted
|
|
||||||
/*
|
|
||||||
func TestMJEG(t *testing.T) {
|
|
||||||
for _, test := range mjpegTests {
|
|
||||||
var buf chunkEncoder
|
|
||||||
err := MJPEG(&buf, bytes.NewReader(test.input), test.delay)
|
|
||||||
if fmt.Sprint(err) != fmt.Sprint(test.err) {
|
|
||||||
t.Errorf("unexpected error for %q: got:%v want:%v", test.name, err, test.err)
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
got := [][]byte(buf)
|
|
||||||
if !reflect.DeepEqual(got, test.want) {
|
|
||||||
t.Errorf("unexpected result for %q:\ngot :%#v\nwant:%#v", test.name, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
*/
|
|
||||||
|
|
||||||
type chunkEncoder [][]byte
|
type chunkEncoder [][]byte
|
||||||
|
|
||||||
func (e *chunkEncoder) Encode(b []byte) error {
|
func (e *chunkEncoder) Encode(b []byte) error {
|
|
@ -0,0 +1,156 @@
|
||||||
|
/*
|
||||||
|
NAME
|
||||||
|
lex.go
|
||||||
|
|
||||||
|
DESCRIPTION
|
||||||
|
lex.go provides a lexer to extract separate JPEG images from a MJPEG stream.
|
||||||
|
|
||||||
|
AUTHOR
|
||||||
|
Dan Kortschak <dan@ausocean.org>
|
||||||
|
|
||||||
|
LICENSE
|
||||||
|
lex.go is Copyright (C) 2017 the Australian Ocean Lab (AusOcean)
|
||||||
|
|
||||||
|
It is free software: you can redistribute it and/or modify them
|
||||||
|
under the terms of the GNU General Public License as published by the
|
||||||
|
Free Software Foundation, either version 3 of the License, or (at your
|
||||||
|
option) any later version.
|
||||||
|
|
||||||
|
It is distributed in the hope that it will be useful, but WITHOUT
|
||||||
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||||
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
||||||
|
for more details.
|
||||||
|
|
||||||
|
You should have received a copy of the GNU General Public License
|
||||||
|
along with revid in gpl.txt. If not, see http://www.gnu.org/licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
// lex.go provides a lexer to extract separate JPEG images from a MJPEG stream.
|
||||||
|
|
||||||
|
package mjpeg
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
var noDelay = make(chan time.Time)
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
close(noDelay)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Lex parses MJPEG frames read from src into separate writes to dst with
|
||||||
|
// successive writes being performed not earlier than the specified delay.
|
||||||
|
func Lex(dst io.Writer, src io.Reader, delay time.Duration) error {
|
||||||
|
var tick <-chan time.Time
|
||||||
|
if delay == 0 {
|
||||||
|
tick = noDelay
|
||||||
|
} else {
|
||||||
|
ticker := time.NewTicker(delay)
|
||||||
|
defer ticker.Stop()
|
||||||
|
tick = ticker.C
|
||||||
|
}
|
||||||
|
|
||||||
|
r := bufio.NewReader(src)
|
||||||
|
for {
|
||||||
|
buf := make([]byte, 2, 4<<10)
|
||||||
|
n, err := r.Read(buf)
|
||||||
|
if n < 2 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if !bytes.Equal(buf, []byte{0xff, 0xd8}) {
|
||||||
|
return fmt.Errorf("parser: not MJPEG frame start: %#v", buf)
|
||||||
|
}
|
||||||
|
var last byte
|
||||||
|
for {
|
||||||
|
b, err := r.ReadByte()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
buf = append(buf, b)
|
||||||
|
if last == 0xff && b == 0xd9 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
last = b
|
||||||
|
}
|
||||||
|
<-tick
|
||||||
|
_, err = dst.Write(buf)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanner is a byte scanner.
|
||||||
|
type scanner struct {
|
||||||
|
buf []byte
|
||||||
|
off int
|
||||||
|
|
||||||
|
// r is the source of data for the scanner.
|
||||||
|
r io.Reader
|
||||||
|
}
|
||||||
|
|
||||||
|
// newScanner returns a scanner initialised with an io.Reader and a read buffer.
|
||||||
|
func newScanner(r io.Reader, buf []byte) *scanner {
|
||||||
|
return &scanner{r: r, buf: buf[:0]}
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanUntilZeroInto scans the scanner's underlying io.Reader until a zero byte
|
||||||
|
// has been read, appending all read bytes to dst. The resulting appended data,
|
||||||
|
// the last read byte and whether the last read byte was zero are returned.
|
||||||
|
func (c *scanner) scanUntilZeroInto(dst []byte) (res []byte, b byte, err error) {
|
||||||
|
outer:
|
||||||
|
for {
|
||||||
|
var i int
|
||||||
|
for i, b = range c.buf[c.off:] {
|
||||||
|
if b != 0x0 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
dst = append(dst, c.buf[c.off:c.off+i+1]...)
|
||||||
|
c.off += i + 1
|
||||||
|
break outer
|
||||||
|
}
|
||||||
|
dst = append(dst, c.buf[c.off:]...)
|
||||||
|
err = c.reload()
|
||||||
|
if err != nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return dst, b, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// readByte is an unexported ReadByte.
|
||||||
|
func (c *scanner) readByte() (byte, error) {
|
||||||
|
if c.off >= len(c.buf) {
|
||||||
|
err := c.reload()
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
b := c.buf[c.off]
|
||||||
|
c.off++
|
||||||
|
return b, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// reload re-fills the scanner's buffer.
|
||||||
|
func (c *scanner) reload() error {
|
||||||
|
n, err := c.r.Read(c.buf[:cap(c.buf)])
|
||||||
|
c.buf = c.buf[:n]
|
||||||
|
if err != nil {
|
||||||
|
if err != io.EOF {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if n == 0 {
|
||||||
|
return io.EOF
|
||||||
|
}
|
||||||
|
}
|
||||||
|
c.off = 0
|
||||||
|
return nil
|
||||||
|
}
|
|
@ -0,0 +1,165 @@
|
||||||
|
/*
|
||||||
|
NAME
|
||||||
|
lex_test.go
|
||||||
|
|
||||||
|
DESCRIPTION
|
||||||
|
lex_test.go provides testing for the lexer in lex.go.
|
||||||
|
|
||||||
|
AUTHOR
|
||||||
|
Dan Kortschak <dan@ausocean.org>
|
||||||
|
|
||||||
|
LICENSE
|
||||||
|
lex_test.go is Copyright (C) 2017 the Australian Ocean Lab (AusOcean)
|
||||||
|
|
||||||
|
It is free software: you can redistribute it and/or modify them
|
||||||
|
under the terms of the GNU General Public License as published by the
|
||||||
|
Free Software Foundation, either version 3 of the License, or (at your
|
||||||
|
option) any later version.
|
||||||
|
|
||||||
|
It is distributed in the hope that it will be useful, but WITHOUT
|
||||||
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||||
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
||||||
|
for more details.
|
||||||
|
|
||||||
|
You should have received a copy of the GNU General Public License
|
||||||
|
along with revid in gpl.txt. If not, see http://www.gnu.org/licenses.
|
||||||
|
*/
|
||||||
|
|
||||||
|
// lex_test.go provides testing for the lexer in lex.go.
|
||||||
|
|
||||||
|
package mjpeg
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"reflect"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
var mjpegTests = []struct {
|
||||||
|
name string
|
||||||
|
input []byte
|
||||||
|
delay time.Duration
|
||||||
|
want [][]byte
|
||||||
|
err error
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "empty",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "null",
|
||||||
|
input: []byte{0xff, 0xd8, 0xff, 0xd9},
|
||||||
|
delay: 0,
|
||||||
|
want: [][]byte{{0xff, 0xd8, 0xff, 0xd9}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "null delayed",
|
||||||
|
input: []byte{0xff, 0xd8, 0xff, 0xd9},
|
||||||
|
delay: time.Millisecond,
|
||||||
|
want: [][]byte{{0xff, 0xd8, 0xff, 0xd9}},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "full",
|
||||||
|
input: []byte{
|
||||||
|
0xff, 0xd8, 'f', 'u', 'l', 'l', 0xff, 0xd9,
|
||||||
|
0xff, 0xd8, 'f', 'r', 'a', 'm', 'e', 0xff, 0xd9,
|
||||||
|
0xff, 0xd8, 'w', 'i', 't', 'h', 0xff, 0xd9,
|
||||||
|
0xff, 0xd8, 'l', 'e', 'n', 'g', 't', 'h', 0xff, 0xd9,
|
||||||
|
0xff, 0xd8, 's', 'p', 'r', 'e', 'a', 'd', 0xff, 0xd9,
|
||||||
|
},
|
||||||
|
delay: 0,
|
||||||
|
want: [][]byte{
|
||||||
|
{0xff, 0xd8, 'f', 'u', 'l', 'l', 0xff, 0xd9},
|
||||||
|
{0xff, 0xd8, 'f', 'r', 'a', 'm', 'e', 0xff, 0xd9},
|
||||||
|
{0xff, 0xd8, 'w', 'i', 't', 'h', 0xff, 0xd9},
|
||||||
|
{0xff, 0xd8, 'l', 'e', 'n', 'g', 't', 'h', 0xff, 0xd9},
|
||||||
|
{0xff, 0xd8, 's', 'p', 'r', 'e', 'a', 'd', 0xff, 0xd9},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "full delayed",
|
||||||
|
input: []byte{
|
||||||
|
0xff, 0xd8, 'f', 'u', 'l', 'l', 0xff, 0xd9,
|
||||||
|
0xff, 0xd8, 'f', 'r', 'a', 'm', 'e', 0xff, 0xd9,
|
||||||
|
0xff, 0xd8, 'w', 'i', 't', 'h', 0xff, 0xd9,
|
||||||
|
0xff, 0xd8, 'l', 'e', 'n', 'g', 't', 'h', 0xff, 0xd9,
|
||||||
|
0xff, 0xd8, 's', 'p', 'r', 'e', 'a', 'd', 0xff, 0xd9,
|
||||||
|
},
|
||||||
|
delay: time.Millisecond,
|
||||||
|
want: [][]byte{
|
||||||
|
{0xff, 0xd8, 'f', 'u', 'l', 'l', 0xff, 0xd9},
|
||||||
|
{0xff, 0xd8, 'f', 'r', 'a', 'm', 'e', 0xff, 0xd9},
|
||||||
|
{0xff, 0xd8, 'w', 'i', 't', 'h', 0xff, 0xd9},
|
||||||
|
{0xff, 0xd8, 'l', 'e', 'n', 'g', 't', 'h', 0xff, 0xd9},
|
||||||
|
{0xff, 0xd8, 's', 'p', 'r', 'e', 'a', 'd', 0xff, 0xd9},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
// FIXME this needs to be adapted
|
||||||
|
/*
|
||||||
|
func Lex(t *testing.T) {
|
||||||
|
for _, test := range mjpegTests {
|
||||||
|
var buf chunkEncoder
|
||||||
|
err := MJPEG(&buf, bytes.NewReader(test.input), test.delay)
|
||||||
|
if fmt.Sprint(err) != fmt.Sprint(test.err) {
|
||||||
|
t.Errorf("unexpected error for %q: got:%v want:%v", test.name, err, test.err)
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
got := [][]byte(buf)
|
||||||
|
if !reflect.DeepEqual(got, test.want) {
|
||||||
|
t.Errorf("unexpected result for %q:\ngot :%#v\nwant:%#v", test.name, got, test.want)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
*/
|
||||||
|
|
||||||
|
type chunkEncoder [][]byte
|
||||||
|
|
||||||
|
func (e *chunkEncoder) Encode(b []byte) error {
|
||||||
|
*e = append(*e, b)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (*chunkEncoder) Stream() <-chan []byte { panic("INVALID USE") }
|
||||||
|
|
||||||
|
func TestScannerReadByte(t *testing.T) {
|
||||||
|
data := []byte("Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua.")
|
||||||
|
|
||||||
|
for _, size := range []int{1, 2, 8, 1 << 10} {
|
||||||
|
r := newScanner(bytes.NewReader(data), make([]byte, size))
|
||||||
|
var got []byte
|
||||||
|
for {
|
||||||
|
b, err := r.readByte()
|
||||||
|
if err != nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
got = append(got, b)
|
||||||
|
}
|
||||||
|
if !bytes.Equal(got, data) {
|
||||||
|
t.Errorf("unexpected result for buffer size %d:\ngot :%q\nwant:%q", size, got, data)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestScannerScanUntilZero(t *testing.T) {
|
||||||
|
data := []byte("Lorem ipsum dolor sit amet, consectetur adipiscing elit,\x00 sed do eiusmod tempor incididunt ut \x00labore et dolore magna aliqua.")
|
||||||
|
|
||||||
|
for _, size := range []int{1, 2, 8, 1 << 10} {
|
||||||
|
r := newScanner(bytes.NewReader(data), make([]byte, size))
|
||||||
|
var got [][]byte
|
||||||
|
for {
|
||||||
|
buf, _, err := r.scanUntilZeroInto(nil)
|
||||||
|
got = append(got, buf)
|
||||||
|
if err != nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
want := bytes.SplitAfter(data, []byte{0})
|
||||||
|
if !reflect.DeepEqual(got, want) {
|
||||||
|
t.Errorf("unexpected result for buffer zie %d:\ngot :%q\nwant:%q", size, got, want)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -38,7 +38,7 @@ import (
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"bitbucket.org/ausocean/av/codec/lex"
|
"bitbucket.org/ausocean/av/codec/h264"
|
||||||
"bitbucket.org/ausocean/av/container/flv"
|
"bitbucket.org/ausocean/av/container/flv"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -199,7 +199,7 @@ func TestFromFrame(t *testing.T) {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Errorf("Failed to create flv encoder with error: %v", err)
|
t.Errorf("Failed to create flv encoder with error: %v", err)
|
||||||
}
|
}
|
||||||
err = lex.H264(flvEncoder, bytes.NewReader(videoData), time.Second/time.Duration(frameRate))
|
err = h264.Lex(flvEncoder, bytes.NewReader(videoData), time.Second/time.Duration(frameRate))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Errorf("Lexing failed with error: %v", err)
|
t.Errorf("Lexing failed with error: %v", err)
|
||||||
}
|
}
|
||||||
|
@ -251,7 +251,7 @@ func TestFromFile(t *testing.T) {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("failed to create encoder: %v", err)
|
t.Fatalf("failed to create encoder: %v", err)
|
||||||
}
|
}
|
||||||
err = lex.H264(flvEncoder, f, time.Second/time.Duration(25))
|
err = h264.Lex(flvEncoder, f, time.Second/time.Duration(25))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Errorf("Lexing and encoding failed with error: %v", err)
|
t.Errorf("Lexing and encoding failed with error: %v", err)
|
||||||
}
|
}
|
||||||
|
|
|
@ -40,7 +40,8 @@ import (
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"bitbucket.org/ausocean/av/codec/lex"
|
"bitbucket.org/ausocean/av/codec/h264"
|
||||||
|
"bitbucket.org/ausocean/av/codec/mjpeg"
|
||||||
"bitbucket.org/ausocean/av/container/flv"
|
"bitbucket.org/ausocean/av/container/flv"
|
||||||
"bitbucket.org/ausocean/av/container/mts"
|
"bitbucket.org/ausocean/av/container/mts"
|
||||||
"bitbucket.org/ausocean/iot/pi/netsender"
|
"bitbucket.org/ausocean/iot/pi/netsender"
|
||||||
|
@ -271,10 +272,10 @@ func (r *Revid) setupPipeline(mtsEnc, flvEnc func(dst io.WriteCloser, rate int)
|
||||||
switch r.config.InputCodec {
|
switch r.config.InputCodec {
|
||||||
case H264:
|
case H264:
|
||||||
r.config.Logger.Log(logger.Info, pkg+"using H264 lexer")
|
r.config.Logger.Log(logger.Info, pkg+"using H264 lexer")
|
||||||
r.lexTo = lex.H264
|
r.lexTo = h264.Lex
|
||||||
case Mjpeg:
|
case Mjpeg:
|
||||||
r.config.Logger.Log(logger.Info, pkg+"using MJPEG lexer")
|
r.config.Logger.Log(logger.Info, pkg+"using MJPEG lexer")
|
||||||
r.lexTo = lex.MJPEG
|
r.lexTo = mjpeg.Lex
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue