blob: d0a4c65de2ee3cf3930089e18ba3cd561e695b57 [file] [log] [blame]
// Copyright 2015 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
// Test broken pipes on Unix systems.
//
//go:build !plan9 && !js
package os_test
import (
"bufio"
"bytes"
"fmt"
"internal/testenv"
"io"
"io/fs"
"os"
"os/exec"
"os/signal"
"runtime"
"strconv"
"strings"
"sync"
"syscall"
"testing"
"time"
)
func TestEPIPE(t *testing.T) {
// This test cannot be run in parallel because of a race similar
// to the one reported in https://go.dev/issue/22315.
//
// Even though the pipe is opened with O_CLOEXEC, if another test forks in
// between the call to os.Pipe and the call to r.Close, that child process can
// retain an open copy of r's file descriptor until it execs. If one of our
// Write calls occurs during that interval it can spuriously succeed,
// buffering the write to the child's copy of the pipe (even though the child
// will not actually read the buffered bytes).
r, w, err := os.Pipe()
if err != nil {
t.Fatal(err)
}
if err := r.Close(); err != nil {
t.Fatal(err)
}
expect := syscall.EPIPE
if runtime.GOOS == "windows" {
// 232 is Windows error code ERROR_NO_DATA, "The pipe is being closed".
expect = syscall.Errno(232)
}
// Every time we write to the pipe we should get an EPIPE.
for i := 0; i < 20; i++ {
_, err = w.Write([]byte("hi"))
if err == nil {
t.Fatal("unexpected success of Write to broken pipe")
}
if pe, ok := err.(*fs.PathError); ok {
err = pe.Err
}
if se, ok := err.(*os.SyscallError); ok {
err = se.Err
}
if err != expect {
t.Errorf("iteration %d: got %v, expected %v", i, err, expect)
}
}
}
func TestStdPipe(t *testing.T) {
switch runtime.GOOS {
case "windows":
t.Skip("Windows doesn't support SIGPIPE")
}
if os.Getenv("GO_TEST_STD_PIPE_HELPER") != "" {
if os.Getenv("GO_TEST_STD_PIPE_HELPER_SIGNAL") != "" {
signal.Notify(make(chan os.Signal, 1), syscall.SIGPIPE)
}
switch os.Getenv("GO_TEST_STD_PIPE_HELPER") {
case "1":
os.Stdout.Write([]byte("stdout"))
case "2":
os.Stderr.Write([]byte("stderr"))
case "3":
if _, err := os.NewFile(3, "3").Write([]byte("3")); err == nil {
os.Exit(3)
}
default:
panic("unrecognized value for GO_TEST_STD_PIPE_HELPER")
}
// For stdout/stderr, we should have crashed with a broken pipe error.
// The caller will be looking for that exit status,
// so just exit normally here to cause a failure in the caller.
// For descriptor 3, a normal exit is expected.
os.Exit(0)
}
testenv.MustHaveExec(t)
// This test cannot be run in parallel due to the same race as for TestEPIPE.
// (We expect a write to a closed pipe can fail, but a concurrent fork of a
// child process can cause the pipe to unexpectedly remain open.)
r, w, err := os.Pipe()
if err != nil {
t.Fatal(err)
}
if err := r.Close(); err != nil {
t.Fatal(err)
}
// Invoke the test program to run the test and write to a closed pipe.
// If sig is false:
// writing to stdout or stderr should cause an immediate SIGPIPE;
// writing to descriptor 3 should fail with EPIPE and then exit 0.
// If sig is true:
// all writes should fail with EPIPE and then exit 0.
for _, sig := range []bool{false, true} {
for dest := 1; dest < 4; dest++ {
cmd := testenv.Command(t, os.Args[0], "-test.run", "TestStdPipe")
cmd.Stdout = w
cmd.Stderr = w
cmd.ExtraFiles = []*os.File{w}
cmd.Env = append(os.Environ(), fmt.Sprintf("GO_TEST_STD_PIPE_HELPER=%d", dest))
if sig {
cmd.Env = append(cmd.Env, "GO_TEST_STD_PIPE_HELPER_SIGNAL=1")
}
if err := cmd.Run(); err == nil {
if !sig && dest < 3 {
t.Errorf("unexpected success of write to closed pipe %d sig %t in child", dest, sig)
}
} else if ee, ok := err.(*exec.ExitError); !ok {
t.Errorf("unexpected exec error type %T: %v", err, err)
} else if ws, ok := ee.Sys().(syscall.WaitStatus); !ok {
t.Errorf("unexpected wait status type %T: %v", ee.Sys(), ee.Sys())
} else if ws.Signaled() && ws.Signal() == syscall.SIGPIPE {
if sig || dest > 2 {
t.Errorf("unexpected SIGPIPE signal for descriptor %d sig %t", dest, sig)
}
} else {
t.Errorf("unexpected exit status %v for descriptor %d sig %t", err, dest, sig)
}
}
}
// Test redirecting stdout but not stderr. Issue 40076.
cmd := testenv.Command(t, os.Args[0], "-test.run", "TestStdPipe")
cmd.Stdout = w
var stderr bytes.Buffer
cmd.Stderr = &stderr
cmd.Env = append(cmd.Environ(), "GO_TEST_STD_PIPE_HELPER=1")
if err := cmd.Run(); err == nil {
t.Errorf("unexpected success of write to closed stdout")
} else if ee, ok := err.(*exec.ExitError); !ok {
t.Errorf("unexpected exec error type %T: %v", err, err)
} else if ws, ok := ee.Sys().(syscall.WaitStatus); !ok {
t.Errorf("unexpected wait status type %T: %v", ee.Sys(), ee.Sys())
} else if !ws.Signaled() || ws.Signal() != syscall.SIGPIPE {
t.Errorf("unexpected exit status %v for write to closed stdout", err)
}
if output := stderr.Bytes(); len(output) > 0 {
t.Errorf("unexpected output on stderr: %s", output)
}
}
func testClosedPipeRace(t *testing.T, read bool) {
// This test cannot be run in parallel due to the same race as for TestEPIPE.
// (We expect a write to a closed pipe can fail, but a concurrent fork of a
// child process can cause the pipe to unexpectedly remain open.)
limit := 1
if !read {
// Get the amount we have to write to overload a pipe
// with no reader.
limit = 131073
if b, err := os.ReadFile("/proc/sys/fs/pipe-max-size"); err == nil {
if i, err := strconv.Atoi(strings.TrimSpace(string(b))); err == nil {
limit = i + 1
}
}
t.Logf("using pipe write limit of %d", limit)
}
r, w, err := os.Pipe()
if err != nil {
t.Fatal(err)
}
defer r.Close()
defer w.Close()
// Close the read end of the pipe in a goroutine while we are
// writing to the write end, or vice-versa.
go func() {
// Give the main goroutine a chance to enter the Read or
// Write call. This is sloppy but the test will pass even
// if we close before the read/write.
time.Sleep(20 * time.Millisecond)
var err error
if read {
err = r.Close()
} else {
err = w.Close()
}
if err != nil {
t.Error(err)
}
}()
b := make([]byte, limit)
if read {
_, err = r.Read(b[:])
} else {
_, err = w.Write(b[:])
}
if err == nil {
t.Error("I/O on closed pipe unexpectedly succeeded")
} else if pe, ok := err.(*fs.PathError); !ok {
t.Errorf("I/O on closed pipe returned unexpected error type %T; expected fs.PathError", pe)
} else if pe.Err != fs.ErrClosed {
t.Errorf("got error %q but expected %q", pe.Err, fs.ErrClosed)
} else {
t.Logf("I/O returned expected error %q", err)
}
}
func TestClosedPipeRaceRead(t *testing.T) {
testClosedPipeRace(t, true)
}
func TestClosedPipeRaceWrite(t *testing.T) {
testClosedPipeRace(t, false)
}
// Issue 20915: Reading on nonblocking fd should not return "waiting
// for unsupported file type." Currently it returns EAGAIN; it is
// possible that in the future it will simply wait for data.
func TestReadNonblockingFd(t *testing.T) {
switch runtime.GOOS {
case "windows":
t.Skip("Windows doesn't support SetNonblock")
}
if os.Getenv("GO_WANT_READ_NONBLOCKING_FD") == "1" {
fd := syscallDescriptor(os.Stdin.Fd())
syscall.SetNonblock(fd, true)
defer syscall.SetNonblock(fd, false)
_, err := os.Stdin.Read(make([]byte, 1))
if err != nil {
if perr, ok := err.(*fs.PathError); !ok || perr.Err != syscall.EAGAIN {
t.Fatalf("read on nonblocking stdin got %q, should have gotten EAGAIN", err)
}
}
os.Exit(0)
}
testenv.MustHaveExec(t)
t.Parallel()
r, w, err := os.Pipe()
if err != nil {
t.Fatal(err)
}
defer r.Close()
defer w.Close()
cmd := testenv.Command(t, os.Args[0], "-test.run="+t.Name())
cmd.Env = append(cmd.Environ(), "GO_WANT_READ_NONBLOCKING_FD=1")
cmd.Stdin = r
output, err := cmd.CombinedOutput()
t.Logf("%s", output)
if err != nil {
t.Errorf("child process failed: %v", err)
}
}
func TestCloseWithBlockingReadByNewFile(t *testing.T) {
t.Parallel()
var p [2]syscallDescriptor
err := syscall.Pipe(p[:])
if err != nil {
t.Fatal(err)
}
// os.NewFile returns a blocking mode file.
testCloseWithBlockingRead(t, os.NewFile(uintptr(p[0]), "reader"), os.NewFile(uintptr(p[1]), "writer"))
}
func TestCloseWithBlockingReadByFd(t *testing.T) {
t.Parallel()
r, w, err := os.Pipe()
if err != nil {
t.Fatal(err)
}
// Calling Fd will put the file into blocking mode.
_ = r.Fd()
testCloseWithBlockingRead(t, r, w)
}
// Test that we don't let a blocking read prevent a close.
func testCloseWithBlockingRead(t *testing.T, r, w *os.File) {
var (
enteringRead = make(chan struct{})
done = make(chan struct{})
)
go func() {
var b [1]byte
close(enteringRead)
_, err := r.Read(b[:])
if err == nil {
t.Error("I/O on closed pipe unexpectedly succeeded")
}
if pe, ok := err.(*fs.PathError); ok {
err = pe.Err
}
if err != io.EOF && err != fs.ErrClosed {
t.Errorf("got %v, expected EOF or closed", err)
}
close(done)
}()
// Give the goroutine a chance to enter the Read
// or Write call. This is sloppy but the test will
// pass even if we close before the read/write.
<-enteringRead
time.Sleep(20 * time.Millisecond)
if err := r.Close(); err != nil {
t.Error(err)
}
// r.Close has completed, but since we assume r is in blocking mode that
// probably didn't unblock the call to r.Read. Close w to unblock it.
w.Close()
<-done
}
func TestPipeEOF(t *testing.T) {
t.Parallel()
r, w, err := os.Pipe()
if err != nil {
t.Fatal(err)
}
testPipeEOF(t, r, w)
}
// testPipeEOF tests that when the write side of a pipe or FIFO is closed,
// a blocked Read call on the reader side returns io.EOF.
//
// This scenario previously failed to unblock the Read call on darwin.
// (See https://go.dev/issue/24164.)
func testPipeEOF(t *testing.T, r io.ReadCloser, w io.WriteCloser) {
// parkDelay is an arbitrary delay we wait for a pipe-reader goroutine to park
// before issuing the corresponding write. The test should pass no matter what
// delay we use, but with a longer delay is has a higher chance of detecting
// poller bugs.
parkDelay := 10 * time.Millisecond
if testing.Short() {
parkDelay = 100 * time.Microsecond
}
writerDone := make(chan struct{})
defer func() {
if err := r.Close(); err != nil {
t.Errorf("error closing reader: %v", err)
}
<-writerDone
}()
write := make(chan int, 1)
go func() {
defer close(writerDone)
for i := range write {
time.Sleep(parkDelay)
_, err := fmt.Fprintf(w, "line %d\n", i)
if err != nil {
t.Errorf("error writing to fifo: %v", err)
return
}
}
time.Sleep(parkDelay)
if err := w.Close(); err != nil {
t.Errorf("error closing writer: %v", err)
}
}()
rbuf := bufio.NewReader(r)
for i := 0; i < 3; i++ {
write <- i
b, err := rbuf.ReadBytes('\n')
if err != nil {
t.Fatal(err)
}
t.Logf("%s\n", bytes.TrimSpace(b))
}
close(write)
b, err := rbuf.ReadBytes('\n')
if err != io.EOF || len(b) != 0 {
t.Errorf(`ReadBytes: %q, %v; want "", io.EOF`, b, err)
}
}
// Issue 24481.
func TestFdRace(t *testing.T) {
// This test starts 100 simultaneous goroutines, which could bury a more
// interesting stack if this or some other test happens to panic. It is also
// nearly instantaneous, so any latency benefit from running it in parallel
// would be minimal.
r, w, err := os.Pipe()
if err != nil {
t.Fatal(err)
}
defer r.Close()
defer w.Close()
var wg sync.WaitGroup
call := func() {
defer wg.Done()
w.Fd()
}
const tries = 100
for i := 0; i < tries; i++ {
wg.Add(1)
go call()
}
wg.Wait()
}
func TestFdReadRace(t *testing.T) {
t.Parallel()
r, w, err := os.Pipe()
if err != nil {
t.Fatal(err)
}
defer r.Close()
defer w.Close()
const count = 10
c := make(chan bool, 1)
var wg sync.WaitGroup
wg.Add(1)
go func() {
defer wg.Done()
var buf [count]byte
r.SetReadDeadline(time.Now().Add(time.Minute))
c <- true
if _, err := r.Read(buf[:]); os.IsTimeout(err) {
t.Error("read timed out")
}
}()
wg.Add(1)
go func() {
defer wg.Done()
<-c
// Give the other goroutine a chance to enter the Read.
// It doesn't matter if this occasionally fails, the test
// will still pass, it just won't test anything.
time.Sleep(10 * time.Millisecond)
r.Fd()
// The bug was that Fd would hang until Read timed out.
// If the bug is fixed, then writing to w and closing r here
// will cause the Read to exit before the timeout expires.
w.Write(make([]byte, count))
r.Close()
}()
wg.Wait()
}