2023-03-30 21:12:00 +00:00
|
|
|
package management
|
|
|
|
|
|
|
|
import (
|
|
|
|
"os"
|
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
jsoniter "github.com/json-iterator/go"
|
|
|
|
"github.com/rs/zerolog"
|
|
|
|
)
|
|
|
|
|
|
|
|
var json = jsoniter.ConfigFastest
|
|
|
|
|
|
|
|
const (
|
|
|
|
// Indicates how many log messages the listener will hold before dropping.
|
|
|
|
// Provides a throttling mechanism to drop latest messages if the sender
|
|
|
|
// can't keep up with the influx of log messages.
|
|
|
|
logWindow = 30
|
|
|
|
)
|
|
|
|
|
|
|
|
// Logger manages the number of management streaming log sessions
|
|
|
|
type Logger struct {
|
|
|
|
sessions []*Session
|
|
|
|
mu sync.RWMutex
|
|
|
|
|
|
|
|
// Unique logger that isn't a io.Writer of the list of zerolog writers. This helps prevent management log
|
|
|
|
// statements from creating infinite recursion to export messages to a session and allows basic debugging and
|
|
|
|
// error statements to be issued in the management code itself.
|
|
|
|
Log *zerolog.Logger
|
|
|
|
}
|
|
|
|
|
|
|
|
func NewLogger() *Logger {
|
|
|
|
log := zerolog.New(zerolog.ConsoleWriter{
|
|
|
|
Out: os.Stdout,
|
|
|
|
TimeFormat: time.RFC3339,
|
2023-04-06 18:31:47 +00:00
|
|
|
}).With().Timestamp().Logger().Level(zerolog.InfoLevel)
|
2023-03-30 21:12:00 +00:00
|
|
|
return &Logger{
|
|
|
|
Log: &log,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
type LoggerListener interface {
|
2023-04-11 16:54:28 +00:00
|
|
|
Listen(*StreamingFilters) *Session
|
2023-03-30 21:12:00 +00:00
|
|
|
Close(*Session)
|
|
|
|
}
|
|
|
|
|
|
|
|
type Session struct {
|
|
|
|
// Buffered channel that holds the recent log events
|
2023-04-06 18:30:42 +00:00
|
|
|
listener chan *Log
|
2023-03-30 21:12:00 +00:00
|
|
|
// Types of log events that this session will provide through the listener
|
2023-04-11 16:54:28 +00:00
|
|
|
filters *StreamingFilters
|
2023-03-30 21:12:00 +00:00
|
|
|
}
|
|
|
|
|
2023-04-11 16:54:28 +00:00
|
|
|
func newSession(size int, filters *StreamingFilters) *Session {
|
|
|
|
s := &Session{
|
2023-04-06 18:30:42 +00:00
|
|
|
listener: make(chan *Log, size),
|
2023-03-30 21:12:00 +00:00
|
|
|
}
|
2023-04-11 16:54:28 +00:00
|
|
|
if filters != nil {
|
|
|
|
s.filters = filters
|
|
|
|
} else {
|
|
|
|
s.filters = &StreamingFilters{}
|
|
|
|
}
|
|
|
|
return s
|
|
|
|
}
|
|
|
|
|
|
|
|
// Insert attempts to insert the log to the session. If the log event matches the provided session filters, it
|
|
|
|
// will be applied to the listener.
|
|
|
|
func (s *Session) Insert(log *Log) {
|
|
|
|
// Level filters are optional
|
|
|
|
if s.filters.Level != nil {
|
|
|
|
if *s.filters.Level > log.Level {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Event filters are optional
|
|
|
|
if len(s.filters.Events) != 0 && !contains(s.filters.Events, log.Event) {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case s.listener <- log:
|
|
|
|
default:
|
|
|
|
// buffer is full, discard
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func contains(array []LogEventType, t LogEventType) bool {
|
|
|
|
for _, v := range array {
|
|
|
|
if v == t {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false
|
2023-03-30 21:12:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Listen creates a new Session that will append filtered log events as they are created.
|
2023-04-11 16:54:28 +00:00
|
|
|
func (l *Logger) Listen(filters *StreamingFilters) *Session {
|
2023-03-30 21:12:00 +00:00
|
|
|
l.mu.Lock()
|
|
|
|
defer l.mu.Unlock()
|
2023-04-11 16:54:28 +00:00
|
|
|
listener := newSession(logWindow, filters)
|
2023-03-30 21:12:00 +00:00
|
|
|
l.sessions = append(l.sessions, listener)
|
|
|
|
return listener
|
|
|
|
}
|
|
|
|
|
|
|
|
// Close will remove a Session from the available sessions that were receiving log events.
|
|
|
|
func (l *Logger) Close(session *Session) {
|
|
|
|
l.mu.Lock()
|
|
|
|
defer l.mu.Unlock()
|
|
|
|
index := -1
|
|
|
|
for i, v := range l.sessions {
|
|
|
|
if v == session {
|
|
|
|
index = i
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if index == -1 {
|
|
|
|
// Not found
|
|
|
|
return
|
|
|
|
}
|
|
|
|
copy(l.sessions[index:], l.sessions[index+1:])
|
|
|
|
l.sessions = l.sessions[:len(l.sessions)-1]
|
|
|
|
}
|
|
|
|
|
|
|
|
// Write will write the log event to all sessions that have available capacity. For those that are full, the message
|
|
|
|
// will be dropped.
|
|
|
|
// This function is the interface that zerolog expects to call when a log event is to be written out.
|
|
|
|
func (l *Logger) Write(p []byte) (int, error) {
|
|
|
|
l.mu.RLock()
|
|
|
|
defer l.mu.RUnlock()
|
|
|
|
// return early if no active sessions
|
|
|
|
if len(l.sessions) == 0 {
|
|
|
|
return len(p), nil
|
|
|
|
}
|
2023-04-06 18:30:42 +00:00
|
|
|
event, err := parseZerologEvent(p)
|
|
|
|
// drop event if unable to parse properly
|
|
|
|
if err != nil {
|
|
|
|
l.Log.Debug().Msg("unable to parse log event")
|
2023-03-30 21:12:00 +00:00
|
|
|
return len(p), nil
|
|
|
|
}
|
2023-04-11 16:54:28 +00:00
|
|
|
for _, session := range l.sessions {
|
|
|
|
session.Insert(event)
|
2023-03-30 21:12:00 +00:00
|
|
|
}
|
|
|
|
return len(p), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (l *Logger) WriteLevel(level zerolog.Level, p []byte) (n int, err error) {
|
|
|
|
return l.Write(p)
|
|
|
|
}
|
2023-04-06 18:30:42 +00:00
|
|
|
|
|
|
|
func parseZerologEvent(p []byte) (*Log, error) {
|
|
|
|
var fields map[string]interface{}
|
|
|
|
iter := json.BorrowIterator(p)
|
|
|
|
defer json.ReturnIterator(iter)
|
|
|
|
iter.ReadVal(&fields)
|
|
|
|
if iter.Error != nil {
|
|
|
|
return nil, iter.Error
|
|
|
|
}
|
|
|
|
logTime := time.Now().UTC().Format(zerolog.TimeFieldFormat)
|
|
|
|
if t, ok := fields[TimeKey]; ok {
|
|
|
|
if t, ok := t.(string); ok {
|
|
|
|
logTime = t
|
|
|
|
}
|
|
|
|
}
|
|
|
|
logLevel := Debug
|
2023-04-11 16:54:28 +00:00
|
|
|
// A zerolog Debug event can be created and then an error can be added
|
|
|
|
// via .Err(error), if so, we upgrade the level to error.
|
|
|
|
if _, hasError := fields["error"]; hasError {
|
|
|
|
logLevel = Error
|
|
|
|
} else {
|
|
|
|
if level, ok := fields[LevelKey]; ok {
|
|
|
|
if level, ok := level.(string); ok {
|
|
|
|
if logLevel, ok = ParseLogLevel(level); !ok {
|
|
|
|
logLevel = Debug
|
|
|
|
}
|
|
|
|
}
|
2023-04-06 18:30:42 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
// Assume the event type is Cloudflared if unable to parse/find. This could be from log events that haven't
|
|
|
|
// yet been tagged with the appropriate EventType yet.
|
|
|
|
logEvent := Cloudflared
|
|
|
|
e := fields[EventTypeKey]
|
|
|
|
if e != nil {
|
|
|
|
if eventNumber, ok := e.(float64); ok {
|
|
|
|
logEvent = LogEventType(eventNumber)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
logMessage := ""
|
|
|
|
if m, ok := fields[MessageKey]; ok {
|
|
|
|
if m, ok := m.(string); ok {
|
|
|
|
logMessage = m
|
|
|
|
}
|
|
|
|
}
|
|
|
|
event := Log{
|
|
|
|
Time: logTime,
|
|
|
|
Level: logLevel,
|
|
|
|
Event: logEvent,
|
|
|
|
Message: logMessage,
|
|
|
|
}
|
|
|
|
// Remove the keys that have top level keys on Log
|
|
|
|
delete(fields, TimeKey)
|
|
|
|
delete(fields, LevelKey)
|
|
|
|
delete(fields, EventTypeKey)
|
|
|
|
delete(fields, MessageKey)
|
|
|
|
// The rest of the keys go into the Fields
|
|
|
|
event.Fields = fields
|
|
|
|
return &event, nil
|
|
|
|
}
|