mirror of https://github.com/tailscale/tailscale/
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
195 lines
5.4 KiB
Go
195 lines
5.4 KiB
Go
5 months ago
|
// Copyright (c) Tailscale Inc & AUTHORS
|
||
|
// SPDX-License-Identifier: BSD-3-Clause
|
||
|
|
||
|
//go:build !plan9
|
||
|
|
||
|
package main
|
||
|
|
||
|
import (
|
||
|
"bytes"
|
||
|
"encoding/binary"
|
||
|
"encoding/json"
|
||
|
"fmt"
|
||
|
"net"
|
||
|
"net/http"
|
||
|
"sync"
|
||
|
"sync/atomic"
|
||
|
|
||
|
"go.uber.org/zap"
|
||
|
corev1 "k8s.io/api/core/v1"
|
||
|
)
|
||
|
|
||
|
// spdyRemoteConnRecorder is a wrapper around net.Conn. It reads the bytestream
|
||
|
// for a 'kubectl exec' session, sends session recording data to the configured
|
||
|
// recorder and forwards the raw bytes to the original destination.
|
||
|
type spdyRemoteConnRecorder struct {
|
||
|
net.Conn
|
||
|
// rec knows how to send data written to it to a tsrecorder instance.
|
||
|
rec *recorder
|
||
|
ch CastHeader
|
||
|
|
||
|
stdoutStreamID atomic.Uint32
|
||
|
stderrStreamID atomic.Uint32
|
||
|
resizeStreamID atomic.Uint32
|
||
|
|
||
|
wmu sync.Mutex // sequences writes
|
||
|
closed bool
|
||
|
failed bool
|
||
|
|
||
|
rmu sync.Mutex // sequences reads
|
||
|
writeCastHeaderOnce sync.Once
|
||
|
|
||
|
zlibReqReader zlibReader
|
||
|
// writeBuf is used to store data written to the connection that has not
|
||
|
// yet been parsed as SPDY frames.
|
||
|
writeBuf bytes.Buffer
|
||
|
// readBuf is used to store data read from the connection that has not
|
||
|
// yet been parsed as SPDY frames.
|
||
|
readBuf bytes.Buffer
|
||
|
log *zap.SugaredLogger
|
||
|
}
|
||
|
|
||
|
// Read reads bytes from the original connection and parses them as SPDY frames.
|
||
|
// If the frame is a data frame for resize stream, sends resize message to the
|
||
|
// recorder. If the frame is a SYN_STREAM control frame that starts stdout,
|
||
|
// stderr or resize stream, store the stream ID.
|
||
|
func (c *spdyRemoteConnRecorder) Read(b []byte) (int, error) {
|
||
|
c.rmu.Lock()
|
||
|
defer c.rmu.Unlock()
|
||
|
n, err := c.Conn.Read(b)
|
||
|
if err != nil {
|
||
|
return n, fmt.Errorf("error reading from connection: %w", err)
|
||
|
}
|
||
|
c.readBuf.Write(b[:n])
|
||
|
|
||
|
var sf spdyFrame
|
||
|
ok, err := sf.Parse(c.readBuf.Bytes(), c.log)
|
||
|
if err != nil {
|
||
|
return 0, fmt.Errorf("error parsing data read from connection: %w", err)
|
||
|
}
|
||
|
if !ok {
|
||
|
// The parsed data in the buffer will be processed together with
|
||
|
// the new data on the next call to Read.
|
||
|
return n, nil
|
||
|
}
|
||
|
c.readBuf.Next(len(sf.Raw)) // advance buffer past the parsed frame
|
||
|
|
||
|
if !sf.Ctrl { // data frame
|
||
|
switch sf.StreamID {
|
||
|
case c.resizeStreamID.Load():
|
||
|
var err error
|
||
|
var msg spdyResizeMsg
|
||
|
if err = json.Unmarshal(sf.Payload, &msg); err != nil {
|
||
|
return 0, fmt.Errorf("error umarshalling resize msg: %w", err)
|
||
|
}
|
||
|
c.ch.Width = msg.Width
|
||
|
c.ch.Height = msg.Height
|
||
|
}
|
||
|
return n, nil
|
||
|
}
|
||
|
// We always want to parse the headers, even if we don't care about the
|
||
|
// frame, as we need to advance the zlib reader otherwise we will get
|
||
|
// garbage.
|
||
|
header, err := sf.parseHeaders(&c.zlibReqReader, c.log)
|
||
|
if err != nil {
|
||
|
return 0, fmt.Errorf("error parsing frame headers: %w", err)
|
||
|
}
|
||
|
if sf.Type == SYN_STREAM {
|
||
|
c.storeStreamID(sf, header)
|
||
|
}
|
||
|
return n, nil
|
||
|
}
|
||
|
|
||
|
// Write forwards the raw data of the latest parsed SPDY frame to the original
|
||
|
// destination. If the frame is an SPDY data frame, it also sends the payload to
|
||
|
// the connected session recorder.
|
||
|
func (c *spdyRemoteConnRecorder) Write(b []byte) (int, error) {
|
||
|
c.wmu.Lock()
|
||
|
defer c.wmu.Unlock()
|
||
|
c.writeBuf.Write(b)
|
||
|
|
||
|
var sf spdyFrame
|
||
|
ok, err := sf.Parse(c.writeBuf.Bytes(), c.log)
|
||
|
if err != nil {
|
||
|
return 0, fmt.Errorf("error parsing data: %w", err)
|
||
|
}
|
||
|
if !ok {
|
||
|
// The parsed data in the buffer will be processed together with
|
||
|
// the new data on the next call to Write.
|
||
|
return len(b), nil
|
||
|
}
|
||
|
c.writeBuf.Next(len(sf.Raw)) // advance buffer past the parsed frame
|
||
|
|
||
|
// If this is a stdout or stderr data frame, send its payload to the
|
||
|
// session recorder.
|
||
|
if !sf.Ctrl {
|
||
|
switch sf.StreamID {
|
||
|
case c.stdoutStreamID.Load(), c.stderrStreamID.Load():
|
||
|
var err error
|
||
|
c.writeCastHeaderOnce.Do(func() {
|
||
|
var j []byte
|
||
|
j, err = json.Marshal(c.ch)
|
||
|
if err != nil {
|
||
|
return
|
||
|
}
|
||
|
j = append(j, '\n')
|
||
|
err = c.rec.writeCastLine(j)
|
||
|
if err != nil {
|
||
|
c.log.Errorf("received error from recorder: %v", err)
|
||
|
}
|
||
|
})
|
||
|
if err != nil {
|
||
|
return 0, fmt.Errorf("error writing CastHeader: %w", err)
|
||
|
}
|
||
|
if err := c.rec.Write(sf.Payload); err != nil {
|
||
|
return 0, fmt.Errorf("error sending payload to session recorder: %w", err)
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
// Forward the whole frame to the original destination.
|
||
|
_, err = c.Conn.Write(sf.Raw) // send to net.Conn
|
||
|
return len(b), err
|
||
|
}
|
||
|
|
||
|
func (c *spdyRemoteConnRecorder) Close() error {
|
||
|
c.wmu.Lock()
|
||
|
defer c.wmu.Unlock()
|
||
|
if c.closed {
|
||
|
return nil
|
||
|
}
|
||
|
if !c.failed && c.writeBuf.Len() > 0 {
|
||
|
c.Conn.Write(c.writeBuf.Bytes())
|
||
|
}
|
||
|
c.writeBuf.Reset()
|
||
|
c.closed = true
|
||
|
err := c.Conn.Close()
|
||
|
c.rec.Close()
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
// parseSynStream parses SYN_STREAM SPDY control frame and updates
|
||
|
// spdyRemoteConnRecorder to store the newly created stream's ID if it is one of
|
||
|
// the stream types we care about. Storing stream_id:stream_type mapping allows
|
||
|
// us to parse received data frames (that have stream IDs) differently depening
|
||
|
// on which stream they belong to (i.e send data frame payload for stdout stream
|
||
|
// to session recorder).
|
||
|
func (c *spdyRemoteConnRecorder) storeStreamID(sf spdyFrame, header http.Header) {
|
||
|
const (
|
||
|
streamTypeHeaderKey = "Streamtype"
|
||
|
)
|
||
|
id := binary.BigEndian.Uint32(sf.Payload[0:4])
|
||
|
switch header.Get(streamTypeHeaderKey) {
|
||
|
case corev1.StreamTypeStdout:
|
||
|
c.stdoutStreamID.Store(id)
|
||
|
case corev1.StreamTypeStderr:
|
||
|
c.stderrStreamID.Store(id)
|
||
|
case corev1.StreamTypeResize:
|
||
|
c.resizeStreamID.Store(id)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
type spdyResizeMsg struct {
|
||
|
Width int `json:"width"`
|
||
|
Height int `json:"height"`
|
||
|
}
|