mirror of
https://github.com/kata-containers/kata-containers.git
synced 2025-09-15 22:08:47 +00:00
logging: Move agent logger to a separate package
The agent logger is useful and generic enough that it can be used by other components, so move the agent logging package to below a top level `pkg` to encourage re-use. Signed-off-by: James O. D. Hunt <james.o.hunt@intel.com>
This commit is contained in:
@@ -6,7 +6,7 @@ edition = "2018"
|
||||
|
||||
[dependencies]
|
||||
oci = { path = "oci" }
|
||||
logging = { path = "logging" }
|
||||
logging = { path = "../../pkg/logging" }
|
||||
rustjail = { path = "rustjail" }
|
||||
protocols = { path = "protocols" }
|
||||
netlink = { path = "netlink" }
|
||||
@@ -33,7 +33,6 @@ tempfile = "3.1.0"
|
||||
|
||||
[workspace]
|
||||
members = [
|
||||
"logging",
|
||||
"netlink",
|
||||
"oci",
|
||||
"protocols",
|
||||
|
@@ -1,20 +0,0 @@
|
||||
[package]
|
||||
name = "logging"
|
||||
version = "0.1.0"
|
||||
authors = ["Tim Zhang <tim@hyper.sh>"]
|
||||
edition = "2018"
|
||||
|
||||
# See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html
|
||||
|
||||
[dependencies]
|
||||
serde_json = "1.0.39"
|
||||
# slog:
|
||||
# - Dynamic keys required to allow HashMap keys to be slog::Serialized.
|
||||
# - The 'max_*' features allow changing the log level at runtime
|
||||
# (by stopping the compiler from removing log calls).
|
||||
slog = { version = "2.5.2", features = ["dynamic-keys", "max_level_trace", "release_max_level_info"] }
|
||||
slog-json = "2.3.0"
|
||||
slog-async = "2.3.0"
|
||||
slog-scope = "4.1.2"
|
||||
# for testing
|
||||
tempfile = "3.1.0"
|
@@ -1,252 +0,0 @@
|
||||
// Copyright (c) 2019 Intel Corporation
|
||||
//
|
||||
// SPDX-License-Identifier: Apache-2.0
|
||||
//
|
||||
#[macro_use]
|
||||
extern crate slog;
|
||||
|
||||
use slog::{BorrowedKV, Drain, Key, OwnedKV, OwnedKVList, Record, KV};
|
||||
use std::collections::HashMap;
|
||||
use std::io;
|
||||
use std::io::Write;
|
||||
use std::process;
|
||||
use std::result;
|
||||
use std::sync::Mutex;
|
||||
|
||||
// XXX: 'writer' param used to make testing possible.
|
||||
pub fn create_logger<W>(name: &str, source: &str, level: slog::Level, writer: W) -> slog::Logger
|
||||
where
|
||||
W: Write + Send + Sync + 'static,
|
||||
{
|
||||
let json_drain = slog_json::Json::new(writer)
|
||||
.add_default_keys()
|
||||
.build()
|
||||
.fuse();
|
||||
|
||||
// Ensure only a unique set of key/value fields is logged
|
||||
let unique_drain = UniqueDrain::new(json_drain).fuse();
|
||||
|
||||
// Allow runtime filtering of records by log level
|
||||
let filter_drain = RuntimeLevelFilter::new(unique_drain, level).fuse();
|
||||
|
||||
// Ensure the logger is thread-safe
|
||||
let async_drain = slog_async::Async::new(filter_drain).build().fuse();
|
||||
|
||||
// Add some "standard" fields
|
||||
slog::Logger::root(
|
||||
async_drain.fuse(),
|
||||
o!("version" => env!("CARGO_PKG_VERSION"),
|
||||
"subsystem" => "root",
|
||||
"pid" => process::id().to_string(),
|
||||
"name" => name.to_string(),
|
||||
"source" => source.to_string()),
|
||||
)
|
||||
}
|
||||
|
||||
// Used to convert an slog::OwnedKVList into a hash map.
|
||||
struct HashSerializer {
|
||||
fields: HashMap<String, String>,
|
||||
}
|
||||
|
||||
impl HashSerializer {
|
||||
fn new() -> HashSerializer {
|
||||
HashSerializer {
|
||||
fields: HashMap::new(),
|
||||
}
|
||||
}
|
||||
|
||||
fn add_field(&mut self, key: String, value: String) {
|
||||
// Take care to only add the first instance of a key. This matters for loggers (but not
|
||||
// Records) since a child loggers have parents and the loggers are serialised child first
|
||||
// meaning the *newest* fields are serialised first.
|
||||
if !self.fields.contains_key(&key) {
|
||||
self.fields.insert(key, value);
|
||||
}
|
||||
}
|
||||
|
||||
fn remove_field(&mut self, key: &str) {
|
||||
self.fields.remove(key);
|
||||
}
|
||||
}
|
||||
|
||||
impl KV for HashSerializer {
|
||||
fn serialize(&self, _record: &Record, serializer: &mut dyn slog::Serializer) -> slog::Result {
|
||||
for (key, value) in self.fields.iter() {
|
||||
serializer.emit_str(Key::from(key.to_string()), value)?;
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
impl slog::Serializer for HashSerializer {
|
||||
fn emit_arguments(&mut self, key: Key, value: &std::fmt::Arguments) -> slog::Result {
|
||||
self.add_field(format!("{}", key), format!("{}", value));
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
struct UniqueDrain<D> {
|
||||
drain: D,
|
||||
}
|
||||
|
||||
impl<D> UniqueDrain<D> {
|
||||
fn new(drain: D) -> Self {
|
||||
UniqueDrain { drain }
|
||||
}
|
||||
}
|
||||
|
||||
impl<D> Drain for UniqueDrain<D>
|
||||
where
|
||||
D: Drain,
|
||||
{
|
||||
type Ok = ();
|
||||
type Err = io::Error;
|
||||
|
||||
fn log(&self, record: &Record, values: &OwnedKVList) -> Result<Self::Ok, Self::Err> {
|
||||
let mut logger_serializer = HashSerializer::new();
|
||||
values.serialize(record, &mut logger_serializer)?;
|
||||
|
||||
let mut record_serializer = HashSerializer::new();
|
||||
record.kv().serialize(record, &mut record_serializer)?;
|
||||
|
||||
for (key, _) in record_serializer.fields.iter() {
|
||||
logger_serializer.remove_field(key);
|
||||
}
|
||||
|
||||
let record_owned_kv = OwnedKV(record_serializer);
|
||||
let record_static = record_static!(record.level(), "");
|
||||
let new_record = Record::new(&record_static, record.msg(), BorrowedKV(&record_owned_kv));
|
||||
|
||||
let logger_owned_kv = OwnedKV(logger_serializer);
|
||||
|
||||
let result = self
|
||||
.drain
|
||||
.log(&new_record, &OwnedKVList::from(logger_owned_kv));
|
||||
|
||||
match result {
|
||||
Ok(_t) => Ok(()),
|
||||
Err(_e) => Err(std::io::Error::new(
|
||||
std::io::ErrorKind::Other,
|
||||
"failed to drain log".to_string(),
|
||||
)),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// A RuntimeLevelFilter will discard all log records whose log level is less than the level
|
||||
// specified in the struct.
|
||||
struct RuntimeLevelFilter<D> {
|
||||
drain: D,
|
||||
level: Mutex<slog::Level>,
|
||||
}
|
||||
|
||||
impl<D> RuntimeLevelFilter<D> {
|
||||
fn new(drain: D, level: slog::Level) -> Self {
|
||||
RuntimeLevelFilter {
|
||||
drain,
|
||||
level: Mutex::new(level),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<D> Drain for RuntimeLevelFilter<D>
|
||||
where
|
||||
D: Drain,
|
||||
{
|
||||
type Ok = Option<D::Ok>;
|
||||
type Err = Option<D::Err>;
|
||||
|
||||
fn log(
|
||||
&self,
|
||||
record: &slog::Record,
|
||||
values: &slog::OwnedKVList,
|
||||
) -> result::Result<Self::Ok, Self::Err> {
|
||||
let log_level = self.level.lock().unwrap();
|
||||
|
||||
if record.level().is_at_least(*log_level) {
|
||||
self.drain.log(record, values)?;
|
||||
}
|
||||
|
||||
Ok(None)
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
use serde_json::Value;
|
||||
use std::io::prelude::*;
|
||||
use tempfile::NamedTempFile;
|
||||
|
||||
#[test]
|
||||
fn test_create_logger_write_to_tmpfile() {
|
||||
// Create a writer for the logger drain to use
|
||||
let writer = NamedTempFile::new().expect("failed to create tempfile");
|
||||
|
||||
// Used to check file contents before the temp file is unlinked
|
||||
let mut writer_ref = writer.reopen().expect("failed to clone tempfile");
|
||||
|
||||
let level = slog::Level::Trace;
|
||||
let name = "name";
|
||||
let source = "source";
|
||||
let record_subsystem = "record-subsystem";
|
||||
|
||||
let record_key = "record-key-1";
|
||||
let record_value = "record-key-2";
|
||||
|
||||
let logger = create_logger(name, source, level, writer);
|
||||
|
||||
let msg = "foo, bar, baz";
|
||||
|
||||
// Call the logger (which calls the drain)
|
||||
info!(logger, "{}", msg; "subsystem" => record_subsystem, record_key => record_value);
|
||||
|
||||
// Force temp file to be flushed
|
||||
drop(logger);
|
||||
|
||||
let mut contents = String::new();
|
||||
writer_ref
|
||||
.read_to_string(&mut contents)
|
||||
.expect("failed to read tempfile contents");
|
||||
|
||||
// Convert file to JSON
|
||||
let fields: Value =
|
||||
serde_json::from_str(&contents).expect("failed to convert logfile to json");
|
||||
|
||||
// Check the expected JSON fields
|
||||
|
||||
let field_ts = fields.get("ts").expect("failed to find timestamp field");
|
||||
assert_ne!(field_ts, "");
|
||||
|
||||
let field_version = fields.get("version").expect("failed to find version field");
|
||||
assert_eq!(field_version, env!("CARGO_PKG_VERSION"));
|
||||
|
||||
let field_pid = fields.get("pid").expect("failed to find pid field");
|
||||
assert_ne!(field_pid, "");
|
||||
|
||||
let field_level = fields.get("level").expect("failed to find level field");
|
||||
assert_eq!(field_level, "INFO");
|
||||
|
||||
let field_msg = fields.get("msg").expect("failed to find msg field");
|
||||
assert_eq!(field_msg, msg);
|
||||
|
||||
let field_name = fields.get("name").expect("failed to find name field");
|
||||
assert_eq!(field_name, name);
|
||||
|
||||
let field_source = fields.get("source").expect("failed to find source field");
|
||||
assert_eq!(field_source, source);
|
||||
|
||||
let field_subsystem = fields
|
||||
.get("subsystem")
|
||||
.expect("failed to find subsystem field");
|
||||
|
||||
// The records field should take priority over the loggers field of the same name
|
||||
assert_eq!(field_subsystem, record_subsystem);
|
||||
|
||||
let field_record_value = fields
|
||||
.get(record_key)
|
||||
.expect("failed to find record key field");
|
||||
assert_eq!(field_record_value, record_value);
|
||||
}
|
||||
}
|
Reference in New Issue
Block a user