mirror of https://bitbucket.org/ausocean/av.git
ring: migrate package to utils/ring
This commit is contained in:
parent
33ef0ce11d
commit
123cf9c7cf
|
@ -46,8 +46,8 @@ import (
|
||||||
|
|
||||||
"bitbucket.org/ausocean/av/generator"
|
"bitbucket.org/ausocean/av/generator"
|
||||||
"bitbucket.org/ausocean/av/parser"
|
"bitbucket.org/ausocean/av/parser"
|
||||||
"bitbucket.org/ausocean/av/ring"
|
|
||||||
"bitbucket.org/ausocean/av/rtmp"
|
"bitbucket.org/ausocean/av/rtmp"
|
||||||
|
"bitbucket.org/ausocean/utils/ring"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Misc constants
|
// Misc constants
|
||||||
|
|
|
@ -1,24 +0,0 @@
|
||||||
# Readme
|
|
||||||
|
|
||||||
Package ring provides a ring buffer of io.ReadWriters.
|
|
||||||
|
|
||||||
# Author
|
|
||||||
|
|
||||||
Dan Kortschak <dan@ausocean.org>
|
|
||||||
|
|
||||||
# License
|
|
||||||
|
|
||||||
ring is Copyright (C) 2017 the Australian Ocean Lab (AusOcean)
|
|
||||||
|
|
||||||
It is free software: you can redistribute it and/or modify them
|
|
||||||
under the terms of the GNU General Public License as published by the
|
|
||||||
Free Software Foundation, either version 3 of the License, or (at your
|
|
||||||
option) any later version.
|
|
||||||
|
|
||||||
It is distributed in the hope that it will be useful, but WITHOUT
|
|
||||||
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
||||||
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
||||||
or more details.
|
|
||||||
|
|
||||||
You should have received a copy of the GNU General Public License
|
|
||||||
along with revid in gpl.txt. If not, see [GNU licenses](http://www.gnu.org/licenses/).
|
|
277
ring/ring.go
277
ring/ring.go
|
@ -1,277 +0,0 @@
|
||||||
/*
|
|
||||||
NAME
|
|
||||||
ring.go - a structure that encapsulates a Buffer datastructure with conccurency
|
|
||||||
functionality
|
|
||||||
|
|
||||||
DESCRIPTION
|
|
||||||
See Readme.md
|
|
||||||
|
|
||||||
AUTHOR
|
|
||||||
Dan Kortschak <dan@ausocean.org>
|
|
||||||
|
|
||||||
LICENSE
|
|
||||||
ring.go is Copyright (C) 2017 the Australian Ocean Lab (AusOcean)
|
|
||||||
|
|
||||||
It is free software: you can redistribute it and/or modify them
|
|
||||||
under the terms of the GNU General Public License as published by the
|
|
||||||
Free Software Foundation, either version 3 of the License, or (at your
|
|
||||||
option) any later version.
|
|
||||||
|
|
||||||
It is distributed in the hope that it will be useful, but WITHOUT
|
|
||||||
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
||||||
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
||||||
for more details.
|
|
||||||
|
|
||||||
You should have received a copy of the GNU General Public License
|
|
||||||
along with revid in gpl.txt. If not, see http://www.gnu.org/licenses.
|
|
||||||
*/
|
|
||||||
|
|
||||||
// Package ring provides a ring buffer of io.ReadWriters.
|
|
||||||
package ring
|
|
||||||
|
|
||||||
import (
|
|
||||||
"errors"
|
|
||||||
"io"
|
|
||||||
"time"
|
|
||||||
)
|
|
||||||
|
|
||||||
var (
|
|
||||||
ErrTimeout = errors.New("ring: buffer cycle timeout")
|
|
||||||
ErrDropped = errors.New("ring: dropped old write")
|
|
||||||
ErrStall = errors.New("ring: unable to dump old write")
|
|
||||||
ErrTooLong = errors.New("ring: write to long for buffer element")
|
|
||||||
)
|
|
||||||
|
|
||||||
// Buffer implements a ring buffer.
|
|
||||||
//
|
|
||||||
// The buffer has a writable head and a readable tail with a queue from the head
|
|
||||||
// to the tail. Concurrent read a write operations are safe.
|
|
||||||
type Buffer struct {
|
|
||||||
head, tail *Chunk
|
|
||||||
full, empty chan *Chunk
|
|
||||||
timeout time.Duration
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewBuffer returns a Buffer with len elements of the given size. The timeout
|
|
||||||
// parameter specifies how long a write operation will wait before failing with
|
|
||||||
// a temporary timeout error.
|
|
||||||
func NewBuffer(len, size int, timeout time.Duration) *Buffer {
|
|
||||||
if len <= 0 || size <= 0 {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
b := Buffer{
|
|
||||||
full: make(chan *Chunk, len),
|
|
||||||
empty: make(chan *Chunk, len),
|
|
||||||
timeout: timeout,
|
|
||||||
}
|
|
||||||
for i := 0; i < len; i++ {
|
|
||||||
b.empty <- newChunk(make([]byte, 0, size))
|
|
||||||
}
|
|
||||||
return &b
|
|
||||||
}
|
|
||||||
|
|
||||||
// Len returns the number of full buffer elements.
|
|
||||||
func (b *Buffer) Len() int {
|
|
||||||
return len(b.full)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Write writes the bytes in b to the next current or next available element of the ring buffer
|
|
||||||
// it returns the number of bytes written and any error.
|
|
||||||
// If no element can be gained within the timeout or stolen from the queue, ErrStall is returned
|
|
||||||
// and if the len(p) is greater than the buffer's element size, ErrTooLong is returned. If a
|
|
||||||
// write was successful but a previous write was dropped, ErrDropped is returned.
|
|
||||||
//
|
|
||||||
// Write is safe to use concurrently with Read, but may not be used concurrently with another
|
|
||||||
// write operation.
|
|
||||||
func (b *Buffer) Write(p []byte) (int, error) {
|
|
||||||
var dropped bool
|
|
||||||
if b.head == nil {
|
|
||||||
timer := time.NewTimer(b.timeout)
|
|
||||||
select {
|
|
||||||
case <-timer.C:
|
|
||||||
select {
|
|
||||||
case b.head = <-b.full:
|
|
||||||
b.head.reset()
|
|
||||||
dropped = true
|
|
||||||
default:
|
|
||||||
// This should never happen.
|
|
||||||
return 0, ErrStall
|
|
||||||
}
|
|
||||||
case b.head = <-b.empty:
|
|
||||||
timer.Stop()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if len(p) > b.head.cap() {
|
|
||||||
return 0, ErrTooLong
|
|
||||||
}
|
|
||||||
if len(p) > b.head.cap()-b.head.Len() {
|
|
||||||
b.full <- b.head
|
|
||||||
b.head = nil
|
|
||||||
return b.Write(p)
|
|
||||||
}
|
|
||||||
n, err := b.head.write(p)
|
|
||||||
if b.head.cap()-b.head.Len() == 0 {
|
|
||||||
b.full <- b.head
|
|
||||||
b.head = nil
|
|
||||||
}
|
|
||||||
if dropped && err == nil {
|
|
||||||
err = ErrDropped
|
|
||||||
}
|
|
||||||
return n, err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Flush puts the currently writing element of the buffer into the queue for reading. Flush
|
|
||||||
// is idempotent.
|
|
||||||
//
|
|
||||||
// Flush is safe to use concurrently with Read, but may not be used concurrently with another
|
|
||||||
// another write operation.
|
|
||||||
func (b *Buffer) Flush() {
|
|
||||||
if b.head == nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
b.full <- b.head
|
|
||||||
b.head = nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close closes the buffer. The buffer may not be written to after a call to close, but can
|
|
||||||
// be drained by calls to Read.
|
|
||||||
//
|
|
||||||
// Flush is safe to use concurrently with Read, but may not be used concurrently with another
|
|
||||||
// another write operation.
|
|
||||||
func (b *Buffer) Close() error {
|
|
||||||
b.Flush()
|
|
||||||
close(b.full)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Next gets the next element from the queue ready for reading, returning ErrTimeout if no
|
|
||||||
// element is available within the timeout. If the Buffer has been closed Next returns io.EOF.
|
|
||||||
//
|
|
||||||
// Is it the responsibility of the caller to close the returned Chunk.
|
|
||||||
//
|
|
||||||
// Next is safe to use concurrently with write operations, but may not be used concurrently with
|
|
||||||
// another Read call or Next call. A goroutine calling Next must not call Flush or Close.
|
|
||||||
func (b *Buffer) Next(timeout time.Duration) (*Chunk, error) {
|
|
||||||
if b.tail == nil {
|
|
||||||
timer := time.NewTimer(timeout)
|
|
||||||
var ok bool
|
|
||||||
select {
|
|
||||||
case <-timer.C:
|
|
||||||
return nil, ErrTimeout
|
|
||||||
case b.tail, ok = <-b.full:
|
|
||||||
timer.Stop()
|
|
||||||
if !ok {
|
|
||||||
return nil, io.EOF
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
b.tail.owner = b
|
|
||||||
return b.tail, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Read reads bytes from the current tail of the ring buffer into p and returns the number of
|
|
||||||
// bytes read and any error.
|
|
||||||
//
|
|
||||||
// Read is safe to use concurrently with write operations, but may not be used concurrently with
|
|
||||||
// another Read call or Next call. A goroutine calling Read must not call Flush or Close.
|
|
||||||
func (b *Buffer) Read(p []byte) (int, error) {
|
|
||||||
if b.tail == nil {
|
|
||||||
return 0, io.EOF
|
|
||||||
}
|
|
||||||
n, err := b.tail.read(p)
|
|
||||||
if b.tail.Len() == 0 {
|
|
||||||
b.tail.reset()
|
|
||||||
b.empty <- b.tail
|
|
||||||
b.tail = nil
|
|
||||||
}
|
|
||||||
return n, err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Chunk is a simplified version of byte buffer without the capacity to grow beyond the
|
|
||||||
// buffer's original cap, and a modified WriteTo method that allows multiple calls without
|
|
||||||
// consuming the buffered data.
|
|
||||||
type Chunk struct {
|
|
||||||
buf []byte
|
|
||||||
off int
|
|
||||||
owner *Buffer
|
|
||||||
}
|
|
||||||
|
|
||||||
func newChunk(buf []byte) *Chunk {
|
|
||||||
return &Chunk{buf: buf[:0]}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Len returns the number of bytes held in the chunk.
|
|
||||||
func (b *Chunk) Len() int {
|
|
||||||
return len(b.buf) - b.off
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *Chunk) cap() int {
|
|
||||||
return cap(b.buf)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *Chunk) reset() {
|
|
||||||
b.buf = b.buf[:0]
|
|
||||||
b.off = 0
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *Chunk) write(p []byte) (n int, err error) {
|
|
||||||
if len(p) > cap(b.buf)-len(b.buf) {
|
|
||||||
err = ErrTooLong
|
|
||||||
}
|
|
||||||
l := len(b.buf)
|
|
||||||
m := l + len(p)
|
|
||||||
if m > cap(b.buf) {
|
|
||||||
m = cap(b.buf)
|
|
||||||
}
|
|
||||||
b.buf = b.buf[:m]
|
|
||||||
n = copy(b.buf[l:], p)
|
|
||||||
return n, err
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *Chunk) read(p []byte) (n int, err error) {
|
|
||||||
if b.Len() <= 0 {
|
|
||||||
if len(p) == 0 {
|
|
||||||
return 0, nil
|
|
||||||
}
|
|
||||||
return 0, io.EOF
|
|
||||||
}
|
|
||||||
n = copy(p, b.buf[b.off:])
|
|
||||||
b.off += n
|
|
||||||
return n, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// WriteTo writes data to w until there's no more data to write or when an error occurs.
|
|
||||||
// The return value n is the number of bytes written. Any error encountered during the
|
|
||||||
// write is also returned. Repeated called to WriteTo will write the same data until
|
|
||||||
// the Chunk's Close method is called.
|
|
||||||
//
|
|
||||||
// WriteTo will panic if the Chunk has not been obtained through a call to Buffer.Next or
|
|
||||||
// has been closed. WriteTo must be used in the same goroutine as the call to Next.
|
|
||||||
func (b *Chunk) WriteTo(w io.Writer) (n int64, err error) {
|
|
||||||
if b.owner == nil || b.owner.tail != b {
|
|
||||||
panic("ring: invalid use of ring buffer chunk")
|
|
||||||
}
|
|
||||||
_n, err := w.Write(b.buf)
|
|
||||||
if _n > len(b.buf) {
|
|
||||||
panic("ring: invalid byte count")
|
|
||||||
}
|
|
||||||
if _n != len(b.buf) {
|
|
||||||
return int64(_n), io.ErrShortWrite
|
|
||||||
}
|
|
||||||
return int64(_n), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Close closes the Chunk, reseting its data and releasing it back to the Buffer. A Chunk
|
|
||||||
// may not be used after it has been closed. Close must be used in the same goroutine as
|
|
||||||
// the call to Next.
|
|
||||||
func (b *Chunk) Close() error {
|
|
||||||
if b.owner == nil || b.owner.tail != b {
|
|
||||||
panic("ring: invalid use of ring buffer chunk")
|
|
||||||
}
|
|
||||||
b.reset()
|
|
||||||
b.owner.tail = nil
|
|
||||||
empty := b.owner.empty
|
|
||||||
b.owner = nil
|
|
||||||
empty <- b
|
|
||||||
return nil
|
|
||||||
}
|
|
|
@ -1,489 +0,0 @@
|
||||||
/*
|
|
||||||
NAME
|
|
||||||
ring_test.go - a test suite adopting the golang testing library to test functionality of the
|
|
||||||
RingBuffer structure
|
|
||||||
|
|
||||||
DESCRIPTION
|
|
||||||
See README.md
|
|
||||||
|
|
||||||
AUTHOR
|
|
||||||
Dan Kortschak <dan@ausocean.org>
|
|
||||||
|
|
||||||
LICENSE
|
|
||||||
ring_test.go is Copyright (C) 2017 the Australian Ocean Lab (AusOcean)
|
|
||||||
|
|
||||||
It is free software: you can redistribute it and/or modify them
|
|
||||||
under the terms of the GNU General Public License as published by the
|
|
||||||
Free Software Foundation, either version 3 of the License, or (at your
|
|
||||||
option) any later version.
|
|
||||||
|
|
||||||
It is distributed in the hope that it will be useful, but WITHOUT
|
|
||||||
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
||||||
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
||||||
for more details.
|
|
||||||
|
|
||||||
You should have received a copy of the GNU General Public License
|
|
||||||
along with revid in gpl.txt. If not, see http://www.gnu.org/licenses.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package ring
|
|
||||||
|
|
||||||
import (
|
|
||||||
"bytes"
|
|
||||||
"io"
|
|
||||||
"io/ioutil"
|
|
||||||
"reflect"
|
|
||||||
"strings"
|
|
||||||
"sync"
|
|
||||||
"testing"
|
|
||||||
"time"
|
|
||||||
)
|
|
||||||
|
|
||||||
var roundTripTests = []struct {
|
|
||||||
name string
|
|
||||||
|
|
||||||
len int
|
|
||||||
size int
|
|
||||||
timeout time.Duration
|
|
||||||
nextTimeout time.Duration
|
|
||||||
|
|
||||||
data [][]string
|
|
||||||
readDelay time.Duration
|
|
||||||
writeDelay time.Duration
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
name: "happy",
|
|
||||||
len: 2, size: 50,
|
|
||||||
timeout: 100 * time.Millisecond,
|
|
||||||
nextTimeout: 100 * time.Millisecond,
|
|
||||||
|
|
||||||
data: [][]string{
|
|
||||||
{"frame1", "frame2", "frame3", "frame4"},
|
|
||||||
{"frame5", "frame6"},
|
|
||||||
{"frame5", "frame6", "frame7"},
|
|
||||||
{"frame8", "frame9", "frame10"},
|
|
||||||
{"frame11"},
|
|
||||||
{"frame12", "frame13"},
|
|
||||||
{"frame14", "frame15", "frame16", "frame17"},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "slow write",
|
|
||||||
len: 2, size: 50,
|
|
||||||
timeout: 100 * time.Millisecond,
|
|
||||||
nextTimeout: 100 * time.Millisecond,
|
|
||||||
|
|
||||||
data: [][]string{
|
|
||||||
{"frame1", "frame2", "frame3", "frame4"},
|
|
||||||
{"frame5", "frame6"},
|
|
||||||
{"frame5", "frame6", "frame7"},
|
|
||||||
{"frame8", "frame9", "frame10"},
|
|
||||||
{"frame11"},
|
|
||||||
{"frame12", "frame13"},
|
|
||||||
{"frame14", "frame15", "frame16", "frame17"},
|
|
||||||
},
|
|
||||||
writeDelay: 500 * time.Millisecond,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "slow read",
|
|
||||||
len: 2, size: 50,
|
|
||||||
timeout: 100 * time.Millisecond,
|
|
||||||
nextTimeout: 100 * time.Millisecond,
|
|
||||||
|
|
||||||
data: [][]string{
|
|
||||||
{"frame1", "frame2", "frame3", "frame4"},
|
|
||||||
{"frame5", "frame6"},
|
|
||||||
{"frame5", "frame6", "frame7"},
|
|
||||||
{"frame8", "frame9", "frame10"},
|
|
||||||
{"frame11"},
|
|
||||||
{"frame12", "frame13"},
|
|
||||||
{"frame14", "frame15", "frame16", "frame17"},
|
|
||||||
},
|
|
||||||
readDelay: 500 * time.Millisecond,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestRoundTrip(t *testing.T) {
|
|
||||||
const maxTimeouts = 100
|
|
||||||
for _, test := range roundTripTests {
|
|
||||||
b := NewBuffer(test.len, test.size, test.timeout)
|
|
||||||
var wg sync.WaitGroup
|
|
||||||
wg.Add(2)
|
|
||||||
go func() {
|
|
||||||
defer wg.Done()
|
|
||||||
for _, c := range test.data {
|
|
||||||
var dropped int
|
|
||||||
for _, f := range c {
|
|
||||||
time.Sleep(test.writeDelay) // Simulate slow data capture.
|
|
||||||
_, err := b.Write([]byte(f))
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
dropped = 0
|
|
||||||
case ErrDropped:
|
|
||||||
if dropped > maxTimeouts {
|
|
||||||
t.Errorf("too many write drops for %q", test.name)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
dropped++
|
|
||||||
default:
|
|
||||||
t.Errorf("unexpected write error for %q: %v", test.name, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
b.Flush()
|
|
||||||
}
|
|
||||||
b.Close()
|
|
||||||
}()
|
|
||||||
go func() {
|
|
||||||
buf := make([]byte, 1<<10)
|
|
||||||
defer wg.Done()
|
|
||||||
var got []string
|
|
||||||
var timeouts int
|
|
||||||
elements:
|
|
||||||
for {
|
|
||||||
_, err := b.Next(test.nextTimeout)
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
timeouts = 0
|
|
||||||
case ErrTimeout:
|
|
||||||
if timeouts > maxTimeouts {
|
|
||||||
t.Errorf("too many timeouts for %q", test.name)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
timeouts++
|
|
||||||
case io.EOF:
|
|
||||||
break elements
|
|
||||||
default:
|
|
||||||
t.Errorf("unexpected read error for %q: %v", test.name, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
reads:
|
|
||||||
for {
|
|
||||||
n, err := b.Read(buf)
|
|
||||||
if n != 0 {
|
|
||||||
time.Sleep(test.readDelay) // Simulate slow data processing.
|
|
||||||
got = append(got, string(buf[:n]))
|
|
||||||
}
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
case io.EOF:
|
|
||||||
break reads
|
|
||||||
default:
|
|
||||||
t.Errorf("unexpected read error for %q: %v", test.name, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
var want []string
|
|
||||||
for _, c := range test.data {
|
|
||||||
want = append(want, strings.Join(c, ""))
|
|
||||||
}
|
|
||||||
if test.readDelay == 0 {
|
|
||||||
if !reflect.DeepEqual(got, want) {
|
|
||||||
t.Errorf("unexpected round-trip result for %q:\ngot: %#v\nwant:%#v", test.name, got, want)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// We may have dropped writes in this case.
|
|
||||||
// So just check that we can consume every
|
|
||||||
// received element with reference to what
|
|
||||||
// was sent.
|
|
||||||
// TODO(kortschak): Check that the number of
|
|
||||||
// missing elements matches the number of
|
|
||||||
// dropped writes.
|
|
||||||
var sidx, ridx int
|
|
||||||
var recd string
|
|
||||||
for ridx, recd = range got {
|
|
||||||
for ; sidx < len(want); sidx++ {
|
|
||||||
if recd == want[sidx] {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if ridx != len(got)-1 {
|
|
||||||
t.Errorf("unexpected round-trip result for %q (unexplained element received):\ngot: %#v\nwant:%#v", test.name, got, want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
wg.Wait()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestRoundTripWriterTo(t *testing.T) {
|
|
||||||
const maxTimeouts = 100
|
|
||||||
for _, test := range roundTripTests {
|
|
||||||
b := NewBuffer(test.len, test.size, test.timeout)
|
|
||||||
var wg sync.WaitGroup
|
|
||||||
wg.Add(2)
|
|
||||||
go func() {
|
|
||||||
defer wg.Done()
|
|
||||||
for _, c := range test.data {
|
|
||||||
var dropped int
|
|
||||||
for _, f := range c {
|
|
||||||
time.Sleep(test.writeDelay) // Simulate slow data capture.
|
|
||||||
_, err := b.Write([]byte(f))
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
dropped = 0
|
|
||||||
case ErrDropped:
|
|
||||||
if dropped > maxTimeouts {
|
|
||||||
t.Errorf("too many write drops for %q", test.name)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
dropped++
|
|
||||||
default:
|
|
||||||
t.Errorf("unexpected write error for %q: %v", test.name, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
b.Flush()
|
|
||||||
}
|
|
||||||
b.Close()
|
|
||||||
}()
|
|
||||||
go func() {
|
|
||||||
var buf bytes.Buffer
|
|
||||||
defer wg.Done()
|
|
||||||
var got []string
|
|
||||||
var timeouts int
|
|
||||||
elements:
|
|
||||||
for {
|
|
||||||
chunk, err := b.Next(test.nextTimeout)
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
timeouts = 0
|
|
||||||
case ErrTimeout:
|
|
||||||
if timeouts > maxTimeouts {
|
|
||||||
t.Errorf("too many timeouts for %q", test.name)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
timeouts++
|
|
||||||
continue
|
|
||||||
case io.EOF:
|
|
||||||
break elements
|
|
||||||
default:
|
|
||||||
t.Errorf("unexpected read error for %q: %v", test.name, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
n, err := chunk.WriteTo(&buf)
|
|
||||||
if n != 0 {
|
|
||||||
time.Sleep(test.readDelay) // Simulate slow data processing.
|
|
||||||
got = append(got, buf.String())
|
|
||||||
buf.Reset()
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("unexpected writeto error for %q: %v", test.name, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
err = chunk.Close()
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("unexpected close error for %q: %v", test.name, err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
var want []string
|
|
||||||
for _, c := range test.data {
|
|
||||||
want = append(want, strings.Join(c, ""))
|
|
||||||
}
|
|
||||||
if test.readDelay == 0 {
|
|
||||||
if !reflect.DeepEqual(got, want) {
|
|
||||||
t.Errorf("unexpected round-trip result for %q:\ngot: %#v\nwant:%#v", test.name, got, want)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// We may have dropped writes in this case.
|
|
||||||
// So just check that we can consume every
|
|
||||||
// received element with reference to what
|
|
||||||
// was sent.
|
|
||||||
// TODO(kortschak): Check that the number of
|
|
||||||
// missing elements matches the number of
|
|
||||||
// dropped writes.
|
|
||||||
var sidx, ridx int
|
|
||||||
var recd string
|
|
||||||
for ridx, recd = range got {
|
|
||||||
for ; sidx < len(want); sidx++ {
|
|
||||||
if recd == want[sidx] {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if ridx != len(got)-1 {
|
|
||||||
t.Errorf("unexpected round-trip result for %q (unexplained element received):\ngot: %#v\nwant:%#v", test.name, got, want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
wg.Wait()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func BenchmarkRoundTrip(b *testing.B) {
|
|
||||||
const (
|
|
||||||
maxTimeouts = 100
|
|
||||||
|
|
||||||
len = 50
|
|
||||||
size = 150e3
|
|
||||||
timeout = 10 * time.Millisecond
|
|
||||||
|
|
||||||
frameLen = 30e3
|
|
||||||
|
|
||||||
writeDelay = 20 * time.Millisecond
|
|
||||||
readDelay = 50 * time.Millisecond
|
|
||||||
)
|
|
||||||
|
|
||||||
// Allocated prior to timer reset since it is an
|
|
||||||
// amortised cost.
|
|
||||||
rb := NewBuffer(len, size, timeout)
|
|
||||||
|
|
||||||
var wg sync.WaitGroup
|
|
||||||
wg.Add(1)
|
|
||||||
go func() {
|
|
||||||
defer wg.Done()
|
|
||||||
var timeouts int
|
|
||||||
elements:
|
|
||||||
for {
|
|
||||||
_, err := rb.Next(timeout)
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
timeouts = 0
|
|
||||||
case ErrTimeout:
|
|
||||||
if timeouts > maxTimeouts {
|
|
||||||
b.Error("too many timeouts")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
timeouts++
|
|
||||||
case io.EOF:
|
|
||||||
break elements
|
|
||||||
default:
|
|
||||||
b.Errorf("unexpected read error: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
_, err = ioutil.ReadAll(rb)
|
|
||||||
time.Sleep(readDelay) // Simulate slow data processing.
|
|
||||||
if err != nil {
|
|
||||||
b.Errorf("unexpected read error: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
data := make([]byte, frameLen)
|
|
||||||
|
|
||||||
b.ResetTimer()
|
|
||||||
b.SetBytes(frameLen)
|
|
||||||
|
|
||||||
var dropped int
|
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
time.Sleep(writeDelay) // Simulate slow data capture.
|
|
||||||
_, err := rb.Write(data)
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
dropped = 0
|
|
||||||
case ErrDropped:
|
|
||||||
if dropped > maxTimeouts {
|
|
||||||
b.Error("too many write drops")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
dropped++
|
|
||||||
default:
|
|
||||||
b.Errorf("unexpected write error: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
rb.Close()
|
|
||||||
|
|
||||||
wg.Wait()
|
|
||||||
}
|
|
||||||
|
|
||||||
func BenchmarkRoundTripWriterTo(b *testing.B) {
|
|
||||||
const (
|
|
||||||
maxTimeouts = 100
|
|
||||||
|
|
||||||
len = 50
|
|
||||||
size = 150e3
|
|
||||||
timeout = 10 * time.Millisecond
|
|
||||||
|
|
||||||
frameLen = 30e3
|
|
||||||
|
|
||||||
writeDelay = 20 * time.Millisecond
|
|
||||||
readDelay = 50 * time.Millisecond
|
|
||||||
)
|
|
||||||
|
|
||||||
// Allocated prior to timer reset since it is an
|
|
||||||
// amortised cost.
|
|
||||||
rb := NewBuffer(len, size, timeout)
|
|
||||||
|
|
||||||
// This is hoisted here to ensure the allocation
|
|
||||||
// is not counted since this is outside the control
|
|
||||||
// of the ring buffer.
|
|
||||||
buf := bytes.NewBuffer(make([]byte, 0, size+1))
|
|
||||||
|
|
||||||
var wg sync.WaitGroup
|
|
||||||
wg.Add(1)
|
|
||||||
go func() {
|
|
||||||
defer wg.Done()
|
|
||||||
var timeouts int
|
|
||||||
elements:
|
|
||||||
for {
|
|
||||||
chunk, err := rb.Next(timeout)
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
timeouts = 0
|
|
||||||
case ErrTimeout:
|
|
||||||
if timeouts > maxTimeouts {
|
|
||||||
b.Error("too many timeouts")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
timeouts++
|
|
||||||
continue
|
|
||||||
case io.EOF:
|
|
||||||
break elements
|
|
||||||
default:
|
|
||||||
b.Errorf("unexpected read error: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
n, err := chunk.WriteTo(buf)
|
|
||||||
if n != 0 {
|
|
||||||
time.Sleep(readDelay) // Simulate slow data processing.
|
|
||||||
buf.Reset()
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
b.Errorf("unexpected writeto error: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
err = chunk.Close()
|
|
||||||
if err != nil {
|
|
||||||
b.Errorf("unexpected close error: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
data := make([]byte, frameLen)
|
|
||||||
|
|
||||||
b.ResetTimer()
|
|
||||||
b.SetBytes(frameLen)
|
|
||||||
|
|
||||||
var dropped int
|
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
time.Sleep(writeDelay) // Simulate slow data capture.
|
|
||||||
_, err := rb.Write(data)
|
|
||||||
switch err {
|
|
||||||
case nil:
|
|
||||||
dropped = 0
|
|
||||||
case ErrDropped:
|
|
||||||
if dropped > maxTimeouts {
|
|
||||||
b.Error("too many write drops")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
dropped++
|
|
||||||
default:
|
|
||||||
b.Errorf("unexpected write error: %v", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
rb.Close()
|
|
||||||
|
|
||||||
wg.Wait()
|
|
||||||
}
|
|
Loading…
Reference in New Issue