From 789be53586425011ea6145c82930cb99ffb4ff31 Mon Sep 17 00:00:00 2001 From: Francesco Romani Date: Tue, 13 May 2025 18:10:54 +0200 Subject: [PATCH 1/5] profilecreator: dissolve helpers into subpackages reorganize the current helper.go dissolving it into specific subpackages. The goal is tstreamline the code layout without changes in behavior. Signed-off-by: Francesco Romani --- cmd/performance-profile-creator/cmd/root.go | 106 ++-------------- .../profilecreator/helper.go | 37 ------ .../profilecreator/profilecreator.go | 10 +- .../profilecreator/profilecreator_test.go | 37 +++--- .../profilecreator/serialize/serialize.go | 120 ++++++++++++++++++ .../profilecreator/toleration/tolerations.go | 9 ++ 6 files changed, 168 insertions(+), 151 deletions(-) delete mode 100644 pkg/performanceprofile/profilecreator/helper.go create mode 100644 pkg/performanceprofile/profilecreator/serialize/serialize.go create mode 100644 pkg/performanceprofile/profilecreator/toleration/tolerations.go diff --git a/cmd/performance-profile-creator/cmd/root.go b/cmd/performance-profile-creator/cmd/root.go index 68c469f417..d21669df15 100644 --- a/cmd/performance-profile-creator/cmd/root.go +++ b/cmd/performance-profile-creator/cmd/root.go @@ -18,9 +18,7 @@ package cmd import ( "bytes" - "encoding/json" "fmt" - "io" "os" "strings" @@ -30,19 +28,19 @@ import ( corev1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" - "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" "k8s.io/apimachinery/pkg/runtime" serializer "k8s.io/apimachinery/pkg/runtime/serializer/json" utilruntime "k8s.io/apimachinery/pkg/util/runtime" "k8s.io/apimachinery/pkg/util/sets" kubeletconfig "k8s.io/kubelet/config/v1beta1" "k8s.io/utils/ptr" - "sigs.k8s.io/yaml" machineconfigv1 "github.com/openshift/api/machineconfiguration/v1" "github.com/openshift/cluster-node-tuning-operator/cmd/performance-profile-creator/cmd/pkg/hypershift" performancev2 "github.com/openshift/cluster-node-tuning-operator/pkg/apis/performanceprofile/v2" "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator" + "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/serialize" + "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/toleration" ) const ( @@ -128,7 +126,7 @@ func NewRootCommand() *cobra.Command { "must-gather-dir-path", } - tolerations := profilecreator.TolerationSet{} + tols := toleration.Set{} root := &cobra.Command{ Use: "performance-profile-creator", @@ -164,7 +162,7 @@ func NewRootCommand() *cobra.Command { return err } - err = profilecreator.EnsureNodesHaveTheSameHardware(nodesHandlers, tolerations) + err = profilecreator.EnsureNodesHaveTheSameHardware(nodesHandlers, tols) if err != nil { return fmt.Errorf("targeted nodes differ: %w", err) } @@ -175,12 +173,17 @@ func NewRootCommand() *cobra.Command { if err != nil { return fmt.Errorf("failed to make profile data from node handler: %w", err) } - tolerations[profilecreator.EnableHardwareTuning] = profileData.enableHardwareTuning + tols[toleration.EnableHardwareTuning] = profileData.enableHardwareTuning profile, err := makePerformanceProfileFrom(*profileData) if err != nil { return err } - return writeProfile(profile, tolerations) + profData, err := serialize.Profile(profile, tols) + if err != nil { + return err + } + _, err = fmt.Println(profData) + return err }, } pcArgs.AddFlags(root.PersistentFlags()) @@ -521,93 +524,6 @@ func makePerformanceProfileFrom(profileData ProfileData) (runtime.Object, error) return profile, nil } -func writeProfile(obj runtime.Object, tolerations profilecreator.TolerationSet) error { - // write CSV to out dir - writer := strings.Builder{} - if err := MarshallObject(obj, &writer); err != nil { - return err - } - - if tolerations[profilecreator.EnableHardwareTuning] { - if _, err := writer.Write([]byte(profilecreator.HardwareTuningMessage)); err != nil { - return err - } - } - - if tolerations[profilecreator.DifferentCoreIDs] { - if _, err := writer.Write([]byte(profilecreator.DifferentCoreIDsMessage)); err != nil { - return err - } - } - - fmt.Printf("%s", writer.String()) - return nil -} - -// MarshallObject mashals an object, usually a CSV into YAML -func MarshallObject(obj interface{}, writer io.Writer) error { - jsonBytes, err := json.Marshal(obj) - if err != nil { - return err - } - - var r unstructured.Unstructured - if err := json.Unmarshal(jsonBytes, &r.Object); err != nil { - return err - } - - // remove status and metadata.creationTimestamp - unstructured.RemoveNestedField(r.Object, "metadata", "creationTimestamp") - unstructured.RemoveNestedField(r.Object, "template", "metadata", "creationTimestamp") - unstructured.RemoveNestedField(r.Object, "spec", "template", "metadata", "creationTimestamp") - unstructured.RemoveNestedField(r.Object, "status") - - deployments, exists, err := unstructured.NestedSlice(r.Object, "spec", "install", "spec", "deployments") - if err != nil { - return err - } - if exists { - for _, obj := range deployments { - deployment := obj.(map[string]interface{}) - unstructured.RemoveNestedField(deployment, "metadata", "creationTimestamp") - unstructured.RemoveNestedField(deployment, "spec", "template", "metadata", "creationTimestamp") - unstructured.RemoveNestedField(deployment, "status") - } - if err := unstructured.SetNestedSlice(r.Object, deployments, "spec", "install", "spec", "deployments"); err != nil { - return err - } - } - - jsonBytes, err = json.Marshal(r.Object) - if err != nil { - return err - } - - yamlBytes, err := yaml.JSONToYAML(jsonBytes) - if err != nil { - return err - } - - // fix double quoted strings by removing unneeded single quotes... - s := string(yamlBytes) - s = strings.Replace(s, " '\"", " \"", -1) - s = strings.Replace(s, "\"'\n", "\"\n", -1) - - yamlBytes = []byte(s) - - _, err = writer.Write([]byte("---\n")) - if err != nil { - return err - } - - _, err = writer.Write(yamlBytes) - if err != nil { - return err - } - - return nil -} - func listNodesForPool(args *ProfileCreatorArgs) ([]*corev1.Node, error) { nodes, err := profilecreator.GetNodeList(args.MustGatherDirPath) if err != nil { diff --git a/pkg/performanceprofile/profilecreator/helper.go b/pkg/performanceprofile/profilecreator/helper.go deleted file mode 100644 index 7f6b69cf77..0000000000 --- a/pkg/performanceprofile/profilecreator/helper.go +++ /dev/null @@ -1,37 +0,0 @@ -package profilecreator - -import ( - v1 "k8s.io/api/core/v1" -) - -const ( - EnableHardwareTuning = "enableHardwareTuning" - HardwareTuningMessage = `# HardwareTuning is an advanced feature, and only intended to be used if -# user is aware of the vendor recommendation on maximum cpu frequency. -# The structure must follow -# -# hardwareTuning: -# isolatedCpuFreq: -# reservedCpuFreq: -` - - DifferentCoreIDs = "differentCoreIDs" - DifferentCoreIDsMessage = `# PPC tolerates having different core IDs for the same logical processors on -# the same NUMA cell compared with other nodes that belong to the stated pool. -# While core IDs numbering may differ between two systems, it still can be considered -# that NUMA and HW topologies are similar; However this depends on the combination -# setting of the hardware, software and firmware as that may affect the mapping pattern. -# While the performance profile controller depends on the logical processors per NUMA, -# having different IDs may affect your system's performance optimization where the cores -# location matters, thus use the generated profile with caution. -` -) - -// TolerationSet records the data to be tolerated or warned about based on the tool handling -type TolerationSet map[string]bool - -func newTestNode(nodeName string) *v1.Node { - n := v1.Node{} - n.Name = nodeName - return &n -} diff --git a/pkg/performanceprofile/profilecreator/profilecreator.go b/pkg/performanceprofile/profilecreator/profilecreator.go index cf0ddd85f5..7780f15aa7 100644 --- a/pkg/performanceprofile/profilecreator/profilecreator.go +++ b/pkg/performanceprofile/profilecreator/profilecreator.go @@ -36,6 +36,8 @@ import ( v1 "k8s.io/api/core/v1" k8syaml "k8s.io/apimachinery/pkg/util/yaml" "k8s.io/utils/cpuset" + + "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/toleration" ) const ( @@ -672,7 +674,7 @@ func (ghwHandler GHWHandler) IsHyperthreadingEnabled() (bool, error) { // EnsureNodesHaveTheSameHardware returns an error if all the input nodes do not have the same hardware configuration and // updates the toleration set to consider as warnings/comments when publishing the generated profile -func EnsureNodesHaveTheSameHardware(nodeHandlers []*GHWHandler, tolerations TolerationSet) error { +func EnsureNodesHaveTheSameHardware(nodeHandlers []*GHWHandler, tols toleration.Set) error { if len(nodeHandlers) < 1 { return fmt.Errorf("no suitable nodes to compare") } @@ -687,7 +689,7 @@ func EnsureNodesHaveTheSameHardware(nodeHandlers []*GHWHandler, tolerations Tole if err != nil { return fmt.Errorf("can't obtain Topology info from GHW snapshot for %s: %v", handle.Node.GetName(), err) } - if err := ensureSameTopology(firstTopology, topology, tolerations); err != nil { + if err := ensureSameTopology(firstTopology, topology, tols); err != nil { return fmt.Errorf("nodes %s and %s have different topology: %v", firstHandle.Node.GetName(), handle.Node.GetName(), err) } } @@ -695,7 +697,7 @@ func EnsureNodesHaveTheSameHardware(nodeHandlers []*GHWHandler, tolerations Tole return nil } -func ensureSameTopology(topology1, topology2 *topology.Info, tolerations TolerationSet) error { +func ensureSameTopology(topology1, topology2 *topology.Info, tols toleration.Set) error { // the assumption here is that both topologies are deep sorted (e.g. slices of numa nodes, cores, processors ..); // see handle.SortedTopology() if topology1.Architecture != topology2.Architecture { @@ -731,7 +733,7 @@ func ensureSameTopology(topology1, topology2 *topology.Info, tolerations Tolerat // core ID equality is treated as best effort. That is because when scheduling workloads, // we care about the logical processors ids and their location on the NUMAs. log.Warnf("the CPU core ids in NUMA node %d differ: %d vs %d", node1.ID, core1.ID, cores2[j].ID) - tolerations[DifferentCoreIDs] = true + tols[toleration.DifferentCoreIDs] = true } if core1.NumThreads != cores2[j].NumThreads { return fmt.Errorf("number of threads for CPU %d in NUMA node %d differs: %d vs %d", core1.ID, node1.ID, core1.NumThreads, cores2[j].NumThreads) diff --git a/pkg/performanceprofile/profilecreator/profilecreator_test.go b/pkg/performanceprofile/profilecreator/profilecreator_test.go index 0d4e1ad557..8bbc99d607 100644 --- a/pkg/performanceprofile/profilecreator/profilecreator_test.go +++ b/pkg/performanceprofile/profilecreator/profilecreator_test.go @@ -20,6 +20,7 @@ import ( log "github.com/sirupsen/logrus" "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/controller/performanceprofile/components" + "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/toleration" ) const ( @@ -1614,10 +1615,10 @@ var _ = Describe("PerformanceProfileCreator: Ensuring Nodes hardware equality", Expect(err).ToNot(HaveOccurred()) nodeHandles := []*GHWHandler{node1Handle, node2Handle} - tols := TolerationSet{} + tols := toleration.Set{} err = EnsureNodesHaveTheSameHardware(nodeHandles, tols) Expect(err).ToNot(HaveOccurred()) - _, ok := tols[DifferentCoreIDs] + _, ok := tols[toleration.DifferentCoreIDs] Expect(ok).To(BeFalse()) }) }) @@ -1638,7 +1639,7 @@ var _ = Describe("PerformanceProfileCreator: Ensuring Nodes hardware equality", Expect(err).ToNot(HaveOccurred()) nodeHandles := []*GHWHandler{node1Handle, node2Handle} - err = EnsureNodesHaveTheSameHardware(nodeHandles, TolerationSet{}) + err = EnsureNodesHaveTheSameHardware(nodeHandles, toleration.Set{}) Expect(err).To(HaveOccurred()) }) }) @@ -1697,50 +1698,50 @@ var _ = Describe("PerformanceProfileCreator: Test Helper Function ensureSameTopo Context("Check if ensureSameTopology is working correctly", func() { It("nodes with similar topology should not return error", func() { - tols := TolerationSet{} + tols := toleration.Set{} err := ensureSameTopology(&originTopology, &mutatedTopology, tols) Expect(err).ToNot(HaveOccurred()) - _, ok := tols[DifferentCoreIDs] + _, ok := tols[toleration.DifferentCoreIDs] Expect(ok).To(BeFalse()) }) It("nodes with different architecture should return error", func() { mutatedTopology.Architecture = topology.ARCHITECTURE_SMP - tols := TolerationSet{} + tols := toleration.Set{} err := ensureSameTopology(&originTopology, &mutatedTopology, tols) Expect(err).To(HaveOccurred()) - _, ok := tols[DifferentCoreIDs] + _, ok := tols[toleration.DifferentCoreIDs] Expect(ok).To(BeFalse()) }) It("nodes with different number of NUMA nodes should return error", func() { mutatedTopology.Nodes = mutatedTopology.Nodes[1:] - tols := TolerationSet{} + tols := toleration.Set{} err := ensureSameTopology(&originTopology, &mutatedTopology, tols) Expect(err).To(HaveOccurred()) - _, ok := tols[DifferentCoreIDs] + _, ok := tols[toleration.DifferentCoreIDs] Expect(ok).To(BeFalse()) }) It("nodes with different number threads per core should return error", func() { mutatedTopology.Nodes[1].Cores[1].NumThreads = 1 - tols := TolerationSet{} + tols := toleration.Set{} err := ensureSameTopology(&originTopology, &mutatedTopology, tols) Expect(err).To(HaveOccurred()) - _, ok := tols[DifferentCoreIDs] + _, ok := tols[toleration.DifferentCoreIDs] Expect(ok).To(BeFalse()) }) It("nodes with different thread IDs should return error", func() { mutatedTopology.Nodes[1].Cores[1].LogicalProcessors[1] = 15 - tols := TolerationSet{} + tols := toleration.Set{} err := ensureSameTopology(&originTopology, &mutatedTopology, tols) Expect(err).To(HaveOccurred()) - _, ok := tols[DifferentCoreIDs] + _, ok := tols[toleration.DifferentCoreIDs] Expect(ok).To(BeFalse()) }) It("same cores with different Core IDs should still considered equivalent but with a warning", func() { mutatedTopology.Nodes[0].Cores[0].ID = 3 - tols := TolerationSet{} + tols := toleration.Set{} err := ensureSameTopology(&originTopology, &mutatedTopology, tols) Expect(err).ToNot(HaveOccurred()) - val, ok := tols[DifferentCoreIDs] + val, ok := tols[toleration.DifferentCoreIDs] Expect(ok).To(BeTrue()) Expect(val).To(BeTrue()) @@ -2003,3 +2004,9 @@ func getSiblingsListForCPUSet(sysinfo systemInfo, cpus cpuset.CPUSet) cpuset.CPU siblingsInt := siblingsSet.UnsortedList() return cpuset.New(siblingsInt...) } + +func newTestNode(nodeName string) *v1.Node { + n := v1.Node{} + n.Name = nodeName + return &n +} diff --git a/pkg/performanceprofile/profilecreator/serialize/serialize.go b/pkg/performanceprofile/profilecreator/serialize/serialize.go new file mode 100644 index 0000000000..6529ebe0d4 --- /dev/null +++ b/pkg/performanceprofile/profilecreator/serialize/serialize.go @@ -0,0 +1,120 @@ +package serialize + +import ( + "encoding/json" + "io" + "strings" + + "k8s.io/apimachinery/pkg/apis/meta/v1/unstructured" + "k8s.io/apimachinery/pkg/runtime" + "sigs.k8s.io/yaml" + + "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/toleration" +) + +const ( + hardwareTuningMessage = `# HardwareTuning is an advanced feature, and only intended to be used if +# user is aware of the vendor recommendation on maximum cpu frequency. +# The structure must follow +# +# hardwareTuning: +# isolatedCpuFreq: +# reservedCpuFreq: +` + + differentCoreIDsMessage = `# PPC tolerates having different core IDs for the same logical processors on +# the same NUMA cell compared with other nodes that belong to the stated pool. +# While core IDs numbering may differ between two systems, it still can be considered +# that NUMA and HW topologies are similar; However this depends on the combination +# setting of the hardware, software and firmware as that may affect the mapping pattern. +# While the performance profile controller depends on the logical processors per NUMA, +# having different IDs may affect your system's performance optimization where the cores +# location matters, thus use the generated profile with caution. +` +) + +func Profile(obj runtime.Object, tols toleration.Set) (string, error) { + // write CSV to out dir + writer := strings.Builder{} + if err := marshallObject(obj, &writer); err != nil { + return "", err + } + + if tols[toleration.EnableHardwareTuning] { + if _, err := writer.Write([]byte(hardwareTuningMessage)); err != nil { + return "", err + } + } + + if tols[toleration.DifferentCoreIDs] { + if _, err := writer.Write([]byte(differentCoreIDsMessage)); err != nil { + return "", err + } + } + + return writer.String(), nil +} + +// MarshallObject marshals an object, usually a CSV into YAML +func marshallObject(obj interface{}, writer io.Writer) error { + jsonBytes, err := json.Marshal(obj) + if err != nil { + return err + } + + var r unstructured.Unstructured + if err := json.Unmarshal(jsonBytes, &r.Object); err != nil { + return err + } + + // remove status and metadata.creationTimestamp + unstructured.RemoveNestedField(r.Object, "metadata", "creationTimestamp") + unstructured.RemoveNestedField(r.Object, "template", "metadata", "creationTimestamp") + unstructured.RemoveNestedField(r.Object, "spec", "template", "metadata", "creationTimestamp") + unstructured.RemoveNestedField(r.Object, "status") + + deployments, exists, err := unstructured.NestedSlice(r.Object, "spec", "install", "spec", "deployments") + if err != nil { + return err + } + if exists { + for _, obj := range deployments { + deployment := obj.(map[string]interface{}) + unstructured.RemoveNestedField(deployment, "metadata", "creationTimestamp") + unstructured.RemoveNestedField(deployment, "spec", "template", "metadata", "creationTimestamp") + unstructured.RemoveNestedField(deployment, "status") + } + if err := unstructured.SetNestedSlice(r.Object, deployments, "spec", "install", "spec", "deployments"); err != nil { + return err + } + } + + jsonBytes, err = json.Marshal(r.Object) + if err != nil { + return err + } + + yamlBytes, err := yaml.JSONToYAML(jsonBytes) + if err != nil { + return err + } + + // fix double quoted strings by removing unneeded single quotes... + s := string(yamlBytes) + s = strings.Replace(s, " '\"", " \"", -1) + s = strings.Replace(s, "\"'\n", "\"\n", -1) + + yamlBytes = []byte(s) + + _, err = writer.Write([]byte("---\n")) + if err != nil { + return err + } + + _, err = writer.Write(yamlBytes) + if err != nil { + return err + } + + return nil +} diff --git a/pkg/performanceprofile/profilecreator/toleration/tolerations.go b/pkg/performanceprofile/profilecreator/toleration/tolerations.go new file mode 100644 index 0000000000..8b2d44fb6d --- /dev/null +++ b/pkg/performanceprofile/profilecreator/toleration/tolerations.go @@ -0,0 +1,9 @@ +package toleration + +const ( + EnableHardwareTuning = "enableHardwareTuning" + DifferentCoreIDs = "differentCoreIDs" +) + +// Set records the data to be tolerated or warned about based on the tool handling +type Set map[string]bool From 2986cb58dc5108b0fd0565a67649b1aebfd31bc8 Mon Sep 17 00:00:00 2001 From: Francesco Romani Date: Tue, 13 May 2025 18:31:56 +0200 Subject: [PATCH 2/5] profilecreator: simplify package layout the `cmd/pkg` layout is awkward. Canonically: - code used only by a tool should sit in `internal/` - code which is/can be exported should sit in `pkg`. However, in case of nontrivial command line utilities, it's a widespread practice to have `pkg/cmd` (or `pkg/commands`) trees. This enable third party utilities to easily embed other utilities maximizing code reuse. We streamline the code layout with trivial code movements - no intended changes of behavior. The net result is now `cmd/performance-profile-creator` is now minified, and holds only the entry point source code file. Signed-off-by: Francesco Romani --- cmd/performance-profile-creator/main.go | 2 +- .../profilecreator/cmd}/hypershift/hypershift.go | 0 .../performanceprofile/profilecreator}/cmd/info.go | 0 .../performanceprofile/profilecreator}/cmd/root.go | 2 +- .../1_performance-profile_creator/ppc.go | 2 +- 5 files changed, 3 insertions(+), 3 deletions(-) rename {cmd/performance-profile-creator/cmd/pkg => pkg/performanceprofile/profilecreator/cmd}/hypershift/hypershift.go (100%) rename {cmd/performance-profile-creator => pkg/performanceprofile/profilecreator}/cmd/info.go (100%) rename {cmd/performance-profile-creator => pkg/performanceprofile/profilecreator}/cmd/root.go (99%) diff --git a/cmd/performance-profile-creator/main.go b/cmd/performance-profile-creator/main.go index 3ba7cdc4d4..61d73330a6 100644 --- a/cmd/performance-profile-creator/main.go +++ b/cmd/performance-profile-creator/main.go @@ -19,7 +19,7 @@ package main import ( "os" - "github.com/openshift/cluster-node-tuning-operator/cmd/performance-profile-creator/cmd" + "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/cmd" ) func main() { diff --git a/cmd/performance-profile-creator/cmd/pkg/hypershift/hypershift.go b/pkg/performanceprofile/profilecreator/cmd/hypershift/hypershift.go similarity index 100% rename from cmd/performance-profile-creator/cmd/pkg/hypershift/hypershift.go rename to pkg/performanceprofile/profilecreator/cmd/hypershift/hypershift.go diff --git a/cmd/performance-profile-creator/cmd/info.go b/pkg/performanceprofile/profilecreator/cmd/info.go similarity index 100% rename from cmd/performance-profile-creator/cmd/info.go rename to pkg/performanceprofile/profilecreator/cmd/info.go diff --git a/cmd/performance-profile-creator/cmd/root.go b/pkg/performanceprofile/profilecreator/cmd/root.go similarity index 99% rename from cmd/performance-profile-creator/cmd/root.go rename to pkg/performanceprofile/profilecreator/cmd/root.go index d21669df15..3ab5831efe 100644 --- a/cmd/performance-profile-creator/cmd/root.go +++ b/pkg/performanceprofile/profilecreator/cmd/root.go @@ -36,9 +36,9 @@ import ( "k8s.io/utils/ptr" machineconfigv1 "github.com/openshift/api/machineconfiguration/v1" - "github.com/openshift/cluster-node-tuning-operator/cmd/performance-profile-creator/cmd/pkg/hypershift" performancev2 "github.com/openshift/cluster-node-tuning-operator/pkg/apis/performanceprofile/v2" "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator" + "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/cmd/hypershift" "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/serialize" "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/toleration" ) diff --git a/test/e2e/performanceprofile/functests-performance-profile-creator/1_performance-profile_creator/ppc.go b/test/e2e/performanceprofile/functests-performance-profile-creator/1_performance-profile_creator/ppc.go index e9c9e09a37..c029fc9043 100644 --- a/test/e2e/performanceprofile/functests-performance-profile-creator/1_performance-profile_creator/ppc.go +++ b/test/e2e/performanceprofile/functests-performance-profile-creator/1_performance-profile_creator/ppc.go @@ -13,8 +13,8 @@ import ( "sigs.k8s.io/yaml" - "github.com/openshift/cluster-node-tuning-operator/cmd/performance-profile-creator/cmd" performancev2 "github.com/openshift/cluster-node-tuning-operator/pkg/apis/performanceprofile/v2" + "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/cmd" testutils "github.com/openshift/cluster-node-tuning-operator/test/e2e/performanceprofile/functests/utils" ) From a0b63c92a3cad4c6513c61411de06eb4dc2bd25b Mon Sep 17 00:00:00 2001 From: Francesco Romani Date: Fri, 4 Jul 2025 11:09:14 +0200 Subject: [PATCH 3/5] profilecreator: chore: refactor profilecreator.go extract code which deals with mustgather or ghw in their own source files. Trivial code movement with no changes in behavior. Signed-off-by: Francesco Romani --- .../profilecreator/ghwhandler.go | 139 +++++++++ .../profilecreator/mustgather.go | 184 ++++++++++++ .../profilecreator/profilecreator.go | 265 ------------------ 3 files changed, 323 insertions(+), 265 deletions(-) create mode 100644 pkg/performanceprofile/profilecreator/ghwhandler.go create mode 100644 pkg/performanceprofile/profilecreator/mustgather.go diff --git a/pkg/performanceprofile/profilecreator/ghwhandler.go b/pkg/performanceprofile/profilecreator/ghwhandler.go new file mode 100644 index 0000000000..07ea9feb8f --- /dev/null +++ b/pkg/performanceprofile/profilecreator/ghwhandler.go @@ -0,0 +1,139 @@ +/* + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + * Copyright 2021 Red Hat, Inc. + */ + +package profilecreator + +import ( + "fmt" + "os" + "path" + "sort" + + "github.com/jaypipes/ghw" + "github.com/jaypipes/ghw/pkg/cpu" + "github.com/jaypipes/ghw/pkg/option" + "github.com/jaypipes/ghw/pkg/topology" + + v1 "k8s.io/api/core/v1" +) + +// NewGHWHandler is a handler to use ghw options corresponding to a node +func NewGHWHandler(mustGatherDirPath string, node *v1.Node) (*GHWHandler, error) { + nodeName := node.GetName() + nodePathSuffix := path.Join(Nodes) + nodepath, err := getMustGatherFullPathsWithFilter(mustGatherDirPath, nodePathSuffix, ClusterScopedResources) + if err != nil { + return nil, fmt.Errorf("can't obtain the node path %s: %v", nodeName, err) + } + _, err = os.Stat(path.Join(nodepath, nodeName, SysInfoFileName)) + if err != nil { + return nil, fmt.Errorf("can't obtain the path: %s for node %s: %v", nodeName, nodepath, err) + } + options := ghw.WithSnapshot(ghw.SnapshotOptions{ + Path: path.Join(nodepath, nodeName, SysInfoFileName), + }) + ghwHandler := &GHWHandler{snapShotOptions: options, Node: node} + return ghwHandler, nil +} + +// GHWHandler is a wrapper around ghw to get the API object +type GHWHandler struct { + snapShotOptions *option.Option + Node *v1.Node +} + +// CPU returns a CPUInfo struct that contains information about the CPUs on the host system +func (ghwHandler GHWHandler) CPU() (*cpu.Info, error) { + return ghw.CPU(ghwHandler.snapShotOptions) +} + +func (ghwHandler GHWHandler) SortedCPU() (*cpu.Info, error) { + cpuInfo, err := ghw.CPU(ghwHandler.snapShotOptions) + if err != nil { + return nil, fmt.Errorf("can't obtain cpuInfo info from GHW snapshot: %v", err) + } + + sort.Slice(cpuInfo.Processors, func(x, y int) bool { + return cpuInfo.Processors[x].ID < cpuInfo.Processors[y].ID + }) + + for _, processor := range cpuInfo.Processors { + for _, core := range processor.Cores { + sort.Slice(core.LogicalProcessors, func(x, y int) bool { + return core.LogicalProcessors[x] < core.LogicalProcessors[y] + }) + } + + sort.Slice(processor.Cores, func(x, y int) bool { + return processor.Cores[x].ID < processor.Cores[y].ID + }) + } + + return cpuInfo, nil +} + +// SortedTopology returns a TopologyInfo struct that contains information about the Topology sorted by numa ids and cpu ids on the host system +func (ghwHandler GHWHandler) SortedTopology() (*topology.Info, error) { + topologyInfo, err := ghw.Topology(ghwHandler.snapShotOptions) + if err != nil { + return nil, fmt.Errorf("can't obtain topology info from GHW snapshot: %v", err) + } + sortTopology(topologyInfo) + return topologyInfo, nil +} + +func sortTopology(topologyInfo *topology.Info) { + sort.Slice(topologyInfo.Nodes, func(x, y int) bool { + return topologyInfo.Nodes[x].ID < topologyInfo.Nodes[y].ID + }) + for _, node := range topologyInfo.Nodes { + for _, core := range node.Cores { + sort.Slice(core.LogicalProcessors, func(x, y int) bool { + return core.LogicalProcessors[x] < core.LogicalProcessors[y] + }) + } + sort.Slice(node.Cores, func(i, j int) bool { + return node.Cores[i].LogicalProcessors[0] < node.Cores[j].LogicalProcessors[0] + }) + } +} + +func (ghwHandler GHWHandler) GatherSystemInfo() (*systemInfo, error) { + cpuInfo, err := ghwHandler.SortedCPU() + if err != nil { + return nil, err + } + + topologyInfo, err := ghwHandler.SortedTopology() + if err != nil { + return nil, err + } + + htEnabled, err := ghwHandler.IsHyperthreadingEnabled() + if err != nil { + return nil, err + } + + return &systemInfo{ + CpuInfo: &extendedCPUInfo{ + CpuInfo: cpuInfo, + NumLogicalProcessorsUsed: make(map[int]int, len(cpuInfo.Processors)), + LogicalProcessorsUsed: make(map[int]struct{}), + }, + TopologyInfo: topologyInfo, + HtEnabled: htEnabled, + }, nil +} diff --git a/pkg/performanceprofile/profilecreator/mustgather.go b/pkg/performanceprofile/profilecreator/mustgather.go new file mode 100644 index 0000000000..29574f5aa1 --- /dev/null +++ b/pkg/performanceprofile/profilecreator/mustgather.go @@ -0,0 +1,184 @@ +/* + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + * + * Copyright 2021 Red Hat, Inc. + */ + +package profilecreator + +import ( + "fmt" + "os" + "path" + "path/filepath" + "strings" + + machineconfigv1 "github.com/openshift/api/machineconfiguration/v1" + log "github.com/sirupsen/logrus" + v1 "k8s.io/api/core/v1" + k8syaml "k8s.io/apimachinery/pkg/util/yaml" +) + +const ( + // ClusterScopedResources defines the subpath, relative to the top-level must-gather directory. + // A top-level must-gather directory is of the following format: + // must-gather-dir/quay-io-openshift-kni-performance-addon-operator-must-gather-sha256- + // Here we find the cluster-scoped definitions saved by must-gather + ClusterScopedResources = "cluster-scoped-resources" + // CoreNodes defines the subpath, relative to ClusterScopedResources, on which we find node-specific data + CoreNodes = "core/nodes" + // MCPools defines the subpath, relative to ClusterScopedResources, on which we find the machine config pool definitions + MCPools = "machineconfiguration.openshift.io/machineconfigpools" + // YAMLSuffix is the extension of the yaml files saved by must-gather + YAMLSuffix = ".yaml" + // Nodes defines the subpath, relative to top-level must-gather directory, on which we find node-specific data + Nodes = "nodes" + // SysInfoFileName defines the name of the file where ghw snapshot is stored + SysInfoFileName = "sysinfo.tgz" + // defines the sub path relative to ClusterScopedResources, on which OCP infrastructure object is + configOCPInfra = "config.openshift.io/infrastructures" +) + +func getMustGatherFullPathsWithFilter(mustGatherPath string, suffix string, filter string) (string, error) { + var paths []string + + // don't assume directory names, only look for the suffix, filter out files having "filter" in their names + err := filepath.Walk(mustGatherPath, func(path string, info os.FileInfo, err error) error { + if strings.HasSuffix(path, suffix) { + if len(filter) == 0 || !strings.Contains(path, filter) { + paths = append(paths, path) + } + } + return nil + }) + if err != nil { + return "", fmt.Errorf("failed to get the path mustGatherPath:%s, suffix:%s %v", mustGatherPath, suffix, err) + } + if len(paths) == 0 { + return "", fmt.Errorf("no match for the specified must gather directory path: %s and suffix: %s", mustGatherPath, suffix) + } + if len(paths) > 1 { + log.Warnf("Multiple matches for the specified must gather directory path: %s and suffix: %s", mustGatherPath, suffix) + return "", fmt.Errorf("Multiple matches for the specified must gather directory path: %s and suffix: %s.\n Expected only one performance-addon-operator-must-gather* directory, please check the must-gather tarball", mustGatherPath, suffix) + } + // returning one possible path + return paths[0], err +} + +func getMustGatherFullPaths(mustGatherPath string, suffix string) (string, error) { + return getMustGatherFullPathsWithFilter(mustGatherPath, suffix, "") +} + +func getNode(mustGatherDirPath, nodeName string) (*v1.Node, error) { + var node v1.Node + nodePathSuffix := path.Join(ClusterScopedResources, CoreNodes, nodeName) + path, err := getMustGatherFullPaths(mustGatherDirPath, nodePathSuffix) + if err != nil { + return nil, fmt.Errorf("failed to get MachineConfigPool for %s: %v", nodeName, err) + } + + src, err := os.Open(path) + if err != nil { + return nil, fmt.Errorf("failed to open %q: %v", path, err) + } + defer src.Close() + + dec := k8syaml.NewYAMLOrJSONDecoder(src, 1024) + if err := dec.Decode(&node); err != nil { + return nil, fmt.Errorf("failed to decode %q: %v", path, err) + } + return &node, nil +} + +// GetNodeList returns the list of nodes using the Node YAMLs stored in Must Gather +func GetNodeList(mustGatherDirPath string) ([]*v1.Node, error) { + machines := make([]*v1.Node, 0) + + nodePathSuffix := path.Join(ClusterScopedResources, CoreNodes) + nodePath, err := getMustGatherFullPaths(mustGatherDirPath, nodePathSuffix) + if err != nil { + return nil, fmt.Errorf("failed to get Nodes from must gather directory: %v", err) + } + if nodePath == "" { + return nil, fmt.Errorf("failed to get Nodes from must gather directory: %v", err) + } + + nodes, err := os.ReadDir(nodePath) + if err != nil { + return nil, fmt.Errorf("failed to list mustGatherPath directories: %v", err) + } + for _, node := range nodes { + nodeName := node.Name() + node, err := getNode(mustGatherDirPath, nodeName) + if err != nil { + return nil, fmt.Errorf("failed to get Nodes %s: %v", nodeName, err) + } + machines = append(machines, node) + } + return machines, nil +} + +// GetMCPList returns the list of MCPs using the mcp YAMLs stored in Must Gather +func GetMCPList(mustGatherDirPath string) ([]*machineconfigv1.MachineConfigPool, error) { + pools := make([]*machineconfigv1.MachineConfigPool, 0) + + mcpPathSuffix := path.Join(ClusterScopedResources, MCPools) + mcpPath, err := getMustGatherFullPaths(mustGatherDirPath, mcpPathSuffix) + if err != nil { + return nil, fmt.Errorf("failed to get MCPs: %v", err) + } + if mcpPath == "" { + return nil, fmt.Errorf("failed to get MCPs path: %v", err) + } + + mcpFiles, err := os.ReadDir(mcpPath) + if err != nil { + return nil, fmt.Errorf("failed to list mustGatherPath directories: %v", err) + } + for _, mcp := range mcpFiles { + mcpName := strings.TrimSuffix(mcp.Name(), filepath.Ext(mcp.Name())) + + mcp, err := GetMCP(mustGatherDirPath, mcpName) + // master pool relevant only when pods can be scheduled on masters, e.g. SNO + if mcpName != "master" && err != nil { + return nil, fmt.Errorf("can't obtain MCP %s: %v", mcpName, err) + } + pools = append(pools, mcp) + } + return pools, nil +} + +// GetMCP returns an MCP object corresponding to a specified MCP Name +func GetMCP(mustGatherDirPath, mcpName string) (*machineconfigv1.MachineConfigPool, error) { + var mcp machineconfigv1.MachineConfigPool + + mcpPathSuffix := path.Join(ClusterScopedResources, MCPools, mcpName+YAMLSuffix) + mcpPath, err := getMustGatherFullPaths(mustGatherDirPath, mcpPathSuffix) + if err != nil { + return nil, fmt.Errorf("failed to obtain MachineConfigPool %s: %v", mcpName, err) + } + if mcpPath == "" { + return nil, fmt.Errorf("failed to obtain MachineConfigPool, mcp:%s does not exist: %v", mcpName, err) + } + + src, err := os.Open(mcpPath) + if err != nil { + return nil, fmt.Errorf("failed to open %q: %v", mcpPath, err) + } + defer src.Close() + dec := k8syaml.NewYAMLOrJSONDecoder(src, 1024) + if err := dec.Decode(&mcp); err != nil { + return nil, fmt.Errorf("failed to decode %q: %v", mcpPath, err) + } + return &mcp, nil +} diff --git a/pkg/performanceprofile/profilecreator/profilecreator.go b/pkg/performanceprofile/profilecreator/profilecreator.go index 7780f15aa7..6fdc8a87ac 100644 --- a/pkg/performanceprofile/profilecreator/profilecreator.go +++ b/pkg/performanceprofile/profilecreator/profilecreator.go @@ -20,20 +20,14 @@ import ( "fmt" "os" "path" - "path/filepath" "reflect" "sort" - "strings" - "github.com/jaypipes/ghw" "github.com/jaypipes/ghw/pkg/cpu" - "github.com/jaypipes/ghw/pkg/option" "github.com/jaypipes/ghw/pkg/topology" log "github.com/sirupsen/logrus" configv1 "github.com/openshift/api/config/v1" - machineconfigv1 "github.com/openshift/api/machineconfiguration/v1" - v1 "k8s.io/api/core/v1" k8syaml "k8s.io/apimachinery/pkg/util/yaml" "k8s.io/utils/cpuset" @@ -41,27 +35,10 @@ import ( ) const ( - // ClusterScopedResources defines the subpath, relative to the top-level must-gather directory. - // A top-level must-gather directory is of the following format: - // must-gather-dir/quay-io-openshift-kni-performance-addon-operator-must-gather-sha256- - // Here we find the cluster-scoped definitions saved by must-gather - ClusterScopedResources = "cluster-scoped-resources" - // CoreNodes defines the subpath, relative to ClusterScopedResources, on which we find node-specific data - CoreNodes = "core/nodes" - // MCPools defines the subpath, relative to ClusterScopedResources, on which we find the machine config pool definitions - MCPools = "machineconfiguration.openshift.io/machineconfigpools" - // YAMLSuffix is the extension of the yaml files saved by must-gather - YAMLSuffix = ".yaml" - // Nodes defines the subpath, relative to top-level must-gather directory, on which we find node-specific data - Nodes = "nodes" - // SysInfoFileName defines the name of the file where ghw snapshot is stored - SysInfoFileName = "sysinfo.tgz" // noSMTKernelArg is the kernel arg value to disable SMT in a system noSMTKernelArg = "nosmt" // allCores correspond to the value when all the processorCores need to be added to the generated CPUset allCores = -1 - // defines the sub path relative to ClusterScopedResources, on which OCP infrastructure object is - configOCPInfra = "config.openshift.io/infrastructures" ) var ( @@ -77,221 +54,6 @@ var ( filterFirstLogicalProcessorInCore = func(index, lpID int) bool { return index != 0 } ) -func getMustGatherFullPathsWithFilter(mustGatherPath string, suffix string, filter string) (string, error) { - var paths []string - - // don't assume directory names, only look for the suffix, filter out files having "filter" in their names - err := filepath.Walk(mustGatherPath, func(path string, info os.FileInfo, err error) error { - if strings.HasSuffix(path, suffix) { - if len(filter) == 0 || !strings.Contains(path, filter) { - paths = append(paths, path) - } - } - return nil - }) - if err != nil { - return "", fmt.Errorf("failed to get the path mustGatherPath:%s, suffix:%s %v", mustGatherPath, suffix, err) - } - if len(paths) == 0 { - return "", fmt.Errorf("no match for the specified must gather directory path: %s and suffix: %s", mustGatherPath, suffix) - } - if len(paths) > 1 { - log.Infof("Multiple matches for the specified must gather directory path: %s and suffix: %s", mustGatherPath, suffix) - return "", fmt.Errorf("Multiple matches for the specified must gather directory path: %s and suffix: %s.\n Expected only one performance-addon-operator-must-gather* directory, please check the must-gather tarball", mustGatherPath, suffix) - } - // returning one possible path - return paths[0], err -} - -func getMustGatherFullPaths(mustGatherPath string, suffix string) (string, error) { - return getMustGatherFullPathsWithFilter(mustGatherPath, suffix, "") -} - -func getNode(mustGatherDirPath, nodeName string) (*v1.Node, error) { - var node v1.Node - nodePathSuffix := path.Join(ClusterScopedResources, CoreNodes, nodeName) - path, err := getMustGatherFullPaths(mustGatherDirPath, nodePathSuffix) - if err != nil { - return nil, fmt.Errorf("failed to get MachineConfigPool for %s: %v", nodeName, err) - } - - src, err := os.Open(path) - if err != nil { - return nil, fmt.Errorf("failed to open %q: %v", path, err) - } - defer src.Close() - - dec := k8syaml.NewYAMLOrJSONDecoder(src, 1024) - if err := dec.Decode(&node); err != nil { - return nil, fmt.Errorf("failed to decode %q: %v", path, err) - } - return &node, nil -} - -// GetNodeList returns the list of nodes using the Node YAMLs stored in Must Gather -func GetNodeList(mustGatherDirPath string) ([]*v1.Node, error) { - machines := make([]*v1.Node, 0) - - nodePathSuffix := path.Join(ClusterScopedResources, CoreNodes) - nodePath, err := getMustGatherFullPaths(mustGatherDirPath, nodePathSuffix) - if err != nil { - return nil, fmt.Errorf("failed to get Nodes from must gather directory: %v", err) - } - if nodePath == "" { - return nil, fmt.Errorf("failed to get Nodes from must gather directory: %v", err) - } - - nodes, err := os.ReadDir(nodePath) - if err != nil { - return nil, fmt.Errorf("failed to list mustGatherPath directories: %v", err) - } - for _, node := range nodes { - nodeName := node.Name() - node, err := getNode(mustGatherDirPath, nodeName) - if err != nil { - return nil, fmt.Errorf("failed to get Nodes %s: %v", nodeName, err) - } - machines = append(machines, node) - } - return machines, nil -} - -// GetMCPList returns the list of MCPs using the mcp YAMLs stored in Must Gather -func GetMCPList(mustGatherDirPath string) ([]*machineconfigv1.MachineConfigPool, error) { - pools := make([]*machineconfigv1.MachineConfigPool, 0) - - mcpPathSuffix := path.Join(ClusterScopedResources, MCPools) - mcpPath, err := getMustGatherFullPaths(mustGatherDirPath, mcpPathSuffix) - if err != nil { - return nil, fmt.Errorf("failed to get MCPs: %v", err) - } - if mcpPath == "" { - return nil, fmt.Errorf("failed to get MCPs path: %v", err) - } - - mcpFiles, err := os.ReadDir(mcpPath) - if err != nil { - return nil, fmt.Errorf("failed to list mustGatherPath directories: %v", err) - } - for _, mcp := range mcpFiles { - mcpName := strings.TrimSuffix(mcp.Name(), filepath.Ext(mcp.Name())) - - mcp, err := GetMCP(mustGatherDirPath, mcpName) - // master pool relevant only when pods can be scheduled on masters, e.g. SNO - if mcpName != "master" && err != nil { - return nil, fmt.Errorf("can't obtain MCP %s: %v", mcpName, err) - } - pools = append(pools, mcp) - } - return pools, nil -} - -// GetMCP returns an MCP object corresponding to a specified MCP Name -func GetMCP(mustGatherDirPath, mcpName string) (*machineconfigv1.MachineConfigPool, error) { - var mcp machineconfigv1.MachineConfigPool - - mcpPathSuffix := path.Join(ClusterScopedResources, MCPools, mcpName+YAMLSuffix) - mcpPath, err := getMustGatherFullPaths(mustGatherDirPath, mcpPathSuffix) - if err != nil { - return nil, fmt.Errorf("failed to obtain MachineConfigPool %s: %v", mcpName, err) - } - if mcpPath == "" { - return nil, fmt.Errorf("failed to obtain MachineConfigPool, mcp:%s does not exist: %v", mcpName, err) - } - - src, err := os.Open(mcpPath) - if err != nil { - return nil, fmt.Errorf("failed to open %q: %v", mcpPath, err) - } - defer src.Close() - dec := k8syaml.NewYAMLOrJSONDecoder(src, 1024) - if err := dec.Decode(&mcp); err != nil { - return nil, fmt.Errorf("failed to decode %q: %v", mcpPath, err) - } - return &mcp, nil -} - -// NewGHWHandler is a handler to use ghw options corresponding to a node -func NewGHWHandler(mustGatherDirPath string, node *v1.Node) (*GHWHandler, error) { - nodeName := node.GetName() - nodePathSuffix := path.Join(Nodes) - nodepath, err := getMustGatherFullPathsWithFilter(mustGatherDirPath, nodePathSuffix, ClusterScopedResources) - if err != nil { - return nil, fmt.Errorf("can't obtain the node path %s: %v", nodeName, err) - } - _, err = os.Stat(path.Join(nodepath, nodeName, SysInfoFileName)) - if err != nil { - return nil, fmt.Errorf("can't obtain the path: %s for node %s: %v", nodeName, nodepath, err) - } - options := ghw.WithSnapshot(ghw.SnapshotOptions{ - Path: path.Join(nodepath, nodeName, SysInfoFileName), - }) - ghwHandler := &GHWHandler{snapShotOptions: options, Node: node} - return ghwHandler, nil -} - -// GHWHandler is a wrapper around ghw to get the API object -type GHWHandler struct { - snapShotOptions *option.Option - Node *v1.Node -} - -// CPU returns a CPUInfo struct that contains information about the CPUs on the host system -func (ghwHandler GHWHandler) CPU() (*cpu.Info, error) { - return ghw.CPU(ghwHandler.snapShotOptions) -} - -func (ghwHandler GHWHandler) SortedCPU() (*cpu.Info, error) { - cpuInfo, err := ghw.CPU(ghwHandler.snapShotOptions) - if err != nil { - return nil, fmt.Errorf("can't obtain cpuInfo info from GHW snapshot: %v", err) - } - - sort.Slice(cpuInfo.Processors, func(x, y int) bool { - return cpuInfo.Processors[x].ID < cpuInfo.Processors[y].ID - }) - - for _, processor := range cpuInfo.Processors { - for _, core := range processor.Cores { - sort.Slice(core.LogicalProcessors, func(x, y int) bool { - return core.LogicalProcessors[x] < core.LogicalProcessors[y] - }) - } - - sort.Slice(processor.Cores, func(x, y int) bool { - return processor.Cores[x].ID < processor.Cores[y].ID - }) - } - - return cpuInfo, nil -} - -// SortedTopology returns a TopologyInfo struct that contains information about the Topology sorted by numa ids and cpu ids on the host system -func (ghwHandler GHWHandler) SortedTopology() (*topology.Info, error) { - topologyInfo, err := ghw.Topology(ghwHandler.snapShotOptions) - if err != nil { - return nil, fmt.Errorf("can't obtain topology info from GHW snapshot: %v", err) - } - sortTopology(topologyInfo) - return topologyInfo, nil -} - -func sortTopology(topologyInfo *topology.Info) { - sort.Slice(topologyInfo.Nodes, func(x, y int) bool { - return topologyInfo.Nodes[x].ID < topologyInfo.Nodes[y].ID - }) - for _, node := range topologyInfo.Nodes { - for _, core := range node.Cores { - sort.Slice(core.LogicalProcessors, func(x, y int) bool { - return core.LogicalProcessors[x] < core.LogicalProcessors[y] - }) - } - sort.Slice(node.Cores, func(i, j int) bool { - return node.Cores[i].LogicalProcessors[0] < node.Cores[j].LogicalProcessors[0] - }) - } -} - // topologyHTDisabled returns topologyinfo in case Hyperthreading needs to be disabled. // It receives a pointer to Topology.Info and deletes logicalprocessors from individual cores. // The behavior of this function depends on ghw data representation. @@ -344,33 +106,6 @@ type systemInfo struct { HtEnabled bool } -func (ghwHandler GHWHandler) GatherSystemInfo() (*systemInfo, error) { - cpuInfo, err := ghwHandler.SortedCPU() - if err != nil { - return nil, err - } - - topologyInfo, err := ghwHandler.SortedTopology() - if err != nil { - return nil, err - } - - htEnabled, err := ghwHandler.IsHyperthreadingEnabled() - if err != nil { - return nil, err - } - - return &systemInfo{ - CpuInfo: &extendedCPUInfo{ - CpuInfo: cpuInfo, - NumLogicalProcessorsUsed: make(map[int]int, len(cpuInfo.Processors)), - LogicalProcessorsUsed: make(map[int]struct{}), - }, - TopologyInfo: topologyInfo, - HtEnabled: htEnabled, - }, nil -} - // Calculates the resevered, isolated and offlined cpuSets. func CalculateCPUSets(systemInfo *systemInfo, reservedCPUCount int, offlinedCPUCount int, splitReservedCPUsAcrossNUMA bool, disableHTFlag bool, highPowerConsumptionMode bool) (cpuset.CPUSet, cpuset.CPUSet, cpuset.CPUSet, error) { topologyInfo := systemInfo.TopologyInfo From 3d91234e6dfeb4d6e6db0a06040d8077998a5f20 Mon Sep 17 00:00:00 2001 From: Francesco Romani Date: Thu, 26 Jun 2025 14:14:30 +0200 Subject: [PATCH 4/5] profilecreator: improvements in the `info` command - add pure text output command. By default (and it's very hard to change, if we ever should) log messages go on stderr. We want the output on stdout. Added flag `--text` to enable it - internal cleanup and code reorganization. In log mode, log all the message in one go, which is easier to manage. No intended change in behavior. Signed-off-by: Francesco Romani --- .../profilecreator/cmd/info.go | 44 ++++++++++++------- 1 file changed, 29 insertions(+), 15 deletions(-) diff --git a/pkg/performanceprofile/profilecreator/cmd/info.go b/pkg/performanceprofile/profilecreator/cmd/info.go index 07570e732d..91a048429e 100644 --- a/pkg/performanceprofile/profilecreator/cmd/info.go +++ b/pkg/performanceprofile/profilecreator/cmd/info.go @@ -5,6 +5,7 @@ import ( "fmt" "os" "sort" + "strings" log "github.com/sirupsen/logrus" "github.com/spf13/cobra" @@ -17,6 +18,7 @@ const ( type infoOptions struct { jsonOutput bool + textOutput bool } // NewInfoCommand return info command, which provides a list of the cluster nodes and their hardware topology. @@ -31,6 +33,7 @@ func NewInfoCommand(pcArgs *ProfileCreatorArgs) *cobra.Command { }, } info.Flags().BoolVar(&opts.jsonOutput, "json", false, "output as JSON") + info.Flags().BoolVar(&opts.textOutput, "text", false, "output as plain text") return info } @@ -40,12 +43,8 @@ func executeInfoMode(mustGatherDirPath string, createForHypershift bool, infoOpt return fmt.Errorf("failed to parse the cluster data: %w", err) } clusterInfo := makeClusterInfoFromClusterData(clusterData) - if infoOpts.jsonOutput { - if err := showClusterInfoJSON(clusterInfo); err != nil { - return fmt.Errorf("unable to show cluster info %w", err) - } - } else { - showClusterInfoLog(clusterInfo) + if err := showClusterInfo(clusterInfo, infoOpts); err != nil { + return fmt.Errorf("unable to show cluster info %w", err) } return nil } @@ -128,21 +127,36 @@ func makeClusterInfoFromClusterData(cluster ClusterData) ClusterInfo { return cInfo.Sort() } -func showClusterInfoJSON(cInfo ClusterInfo) error { - return json.NewEncoder(os.Stdout).Encode(cInfo) +func showClusterInfo(cInfo ClusterInfo, infoOpts *infoOptions) error { + if infoOpts.jsonOutput { + return json.NewEncoder(os.Stdout).Encode(cInfo) + } + textInfo := dumpClusterInfo(cInfo) + if infoOpts.textOutput { + fmt.Print(textInfo) + return nil + } + log.Infof("Cluster info:\n%s", textInfo) + return nil } -func showClusterInfoLog(cInfo ClusterInfo) { - log.Infof("Cluster info:") +func dumpClusterInfo(cInfo ClusterInfo) string { + var sb strings.Builder for _, mcpInfo := range cInfo { - log.Infof("MCP '%s' nodes:", mcpInfo.Name) + fmt.Fprintf(&sb, "MCP '%s' nodes:", mcpInfo.Name) + sb.WriteString("\n") for _, nInfo := range mcpInfo.Nodes { - log.Infof("Node: %s (NUMA cells: %d, HT: %v)", nInfo.Name, len(nInfo.NUMACells), nInfo.HTEnabled) + fmt.Fprintf(&sb, "Node: %s (NUMA cells: %d, HT: %v)", nInfo.Name, len(nInfo.NUMACells), nInfo.HTEnabled) + sb.WriteString("\n") for _, cInfo := range nInfo.NUMACells { - log.Infof("NUMA cell %d : %v", cInfo.ID, cInfo.CoreList) + fmt.Fprintf(&sb, "NUMA cell %d : %v", cInfo.ID, cInfo.CoreList) + sb.WriteString("\n") } - log.Infof("CPU(s): %d", nInfo.CPUsCount) + fmt.Fprintf(&sb, "CPU(s): %d", nInfo.CPUsCount) + sb.WriteString("\n") } - log.Infof("---") + fmt.Fprintf(&sb, "---") + sb.WriteString("\n") } + return sb.String() } From 405c966162f3860440ebb05ffb78527792161d60 Mon Sep 17 00:00:00 2001 From: Francesco Romani Date: Thu, 26 Jun 2025 14:45:25 +0200 Subject: [PATCH 5/5] profilecreator: add and use Alert function We used the `logrus` package internally, but minimally and pretty much only to notify processing status to the user. This code was thus the only remaining consumer of that package, which we don't really need. Add a neutral (nor log, nor warning) minimal abstraction `Alert`, similar in spirit to `ghw`, to be used when we need to notify the user about processing status. Make the backend easy to replace to make integration of this code easier. We don't see this happening, but the extra cost is negligible. Remove the direct dependency to `logrus`. Note we still have quite many indirect dependencies pulling it in. Signed-off-by: Francesco Romani --- go.mod | 1 - go.sum | 3 - .../profilecreator/alert.go | 30 + .../profilecreator/cmd/info.go | 7 +- .../profilecreator/cmd/root.go | 14 +- pkg/performanceprofile/profilecreator/mcp.go | 6 +- .../profilecreator/mustgather.go | 3 +- .../profilecreator/profilecreator.go | 19 +- .../profilecreator/profilecreator_test.go | 164 +++--- vendor/github.com/sirupsen/logrus/.gitignore | 4 - .../github.com/sirupsen/logrus/.golangci.yml | 40 -- vendor/github.com/sirupsen/logrus/.travis.yml | 15 - .../github.com/sirupsen/logrus/CHANGELOG.md | 259 --------- vendor/github.com/sirupsen/logrus/LICENSE | 21 - vendor/github.com/sirupsen/logrus/README.md | 515 ------------------ vendor/github.com/sirupsen/logrus/alt_exit.go | 76 --- .../github.com/sirupsen/logrus/appveyor.yml | 14 - .../github.com/sirupsen/logrus/buffer_pool.go | 43 -- vendor/github.com/sirupsen/logrus/doc.go | 26 - vendor/github.com/sirupsen/logrus/entry.go | 442 --------------- vendor/github.com/sirupsen/logrus/exported.go | 270 --------- .../github.com/sirupsen/logrus/formatter.go | 78 --- vendor/github.com/sirupsen/logrus/hooks.go | 34 -- .../sirupsen/logrus/json_formatter.go | 128 ----- vendor/github.com/sirupsen/logrus/logger.go | 417 -------------- vendor/github.com/sirupsen/logrus/logrus.go | 186 ------- .../logrus/terminal_check_appengine.go | 11 - .../sirupsen/logrus/terminal_check_bsd.go | 13 - .../sirupsen/logrus/terminal_check_js.go | 7 - .../logrus/terminal_check_no_terminal.go | 11 - .../logrus/terminal_check_notappengine.go | 17 - .../sirupsen/logrus/terminal_check_solaris.go | 11 - .../sirupsen/logrus/terminal_check_unix.go | 13 - .../sirupsen/logrus/terminal_check_windows.go | 27 - .../sirupsen/logrus/text_formatter.go | 339 ------------ vendor/github.com/sirupsen/logrus/writer.go | 102 ---- vendor/modules.txt | 3 - 37 files changed, 133 insertions(+), 3236 deletions(-) create mode 100644 pkg/performanceprofile/profilecreator/alert.go delete mode 100644 vendor/github.com/sirupsen/logrus/.gitignore delete mode 100644 vendor/github.com/sirupsen/logrus/.golangci.yml delete mode 100644 vendor/github.com/sirupsen/logrus/.travis.yml delete mode 100644 vendor/github.com/sirupsen/logrus/CHANGELOG.md delete mode 100644 vendor/github.com/sirupsen/logrus/LICENSE delete mode 100644 vendor/github.com/sirupsen/logrus/README.md delete mode 100644 vendor/github.com/sirupsen/logrus/alt_exit.go delete mode 100644 vendor/github.com/sirupsen/logrus/appveyor.yml delete mode 100644 vendor/github.com/sirupsen/logrus/buffer_pool.go delete mode 100644 vendor/github.com/sirupsen/logrus/doc.go delete mode 100644 vendor/github.com/sirupsen/logrus/entry.go delete mode 100644 vendor/github.com/sirupsen/logrus/exported.go delete mode 100644 vendor/github.com/sirupsen/logrus/formatter.go delete mode 100644 vendor/github.com/sirupsen/logrus/hooks.go delete mode 100644 vendor/github.com/sirupsen/logrus/json_formatter.go delete mode 100644 vendor/github.com/sirupsen/logrus/logger.go delete mode 100644 vendor/github.com/sirupsen/logrus/logrus.go delete mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_appengine.go delete mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_bsd.go delete mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_js.go delete mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_no_terminal.go delete mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_notappengine.go delete mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_solaris.go delete mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_unix.go delete mode 100644 vendor/github.com/sirupsen/logrus/terminal_check_windows.go delete mode 100644 vendor/github.com/sirupsen/logrus/text_formatter.go delete mode 100644 vendor/github.com/sirupsen/logrus/writer.go diff --git a/go.mod b/go.mod index 7856a4650c..2c289bb96b 100644 --- a/go.mod +++ b/go.mod @@ -26,7 +26,6 @@ require ( github.com/openshift/library-go v0.0.0-20240419113445-f1541d628746 github.com/pkg/errors v0.9.1 github.com/prometheus/client_golang v1.21.1 - github.com/sirupsen/logrus v1.9.3 github.com/spf13/cobra v1.9.1 github.com/spf13/pflag v1.0.6 gopkg.in/fsnotify.v1 v1.4.7 diff --git a/go.sum b/go.sum index a6cecef60f..ea705f8a4a 100644 --- a/go.sum +++ b/go.sum @@ -434,8 +434,6 @@ github.com/rwcarlsen/goexif v0.0.0-20190401172101-9e8deecbddbd/go.mod h1:hPqNNc0 github.com/sirupsen/logrus v1.4.0/go.mod h1:LxeOpSwHxABJmUn/MG1IvRgCAasNZTLOkJPxbbu5VWo= github.com/sirupsen/logrus v1.4.1/go.mod h1:ni0Sbl8bgC9z8RoU9G6nDWqqs/fq4eDPysMBDgk/93Q= github.com/sirupsen/logrus v1.4.2/go.mod h1:tLMulIdttU9McNUspp0xgXVQah82FyeX6MwdIuYE2rE= -github.com/sirupsen/logrus v1.9.3 h1:dueUQJ1C2q9oE3F7wvmSGAaVtTmUizReu6fjN8uqzbQ= -github.com/sirupsen/logrus v1.9.3/go.mod h1:naHLuLoDiP4jHNo9R0sCBMtWGeIprob74mVsIT4qYEQ= github.com/spf13/cobra v0.0.3/go.mod h1:1l0Ry5zgKvJasoi3XT1TypsSe7PqH0Sj9dhYf7v3XqQ= github.com/spf13/cobra v1.9.1 h1:CXSaggrXdbHK9CF+8ywj8Amf7PBRmPCOJugH954Nnlo= github.com/spf13/cobra v1.9.1/go.mod h1:nDyEzZ8ogv936Cinf6g1RU9MRY64Ir93oCnqb9wxYW0= @@ -686,7 +684,6 @@ golang.org/x/sys v0.0.0-20220310020820-b874c991c1a5/go.mod h1:oPkhp1MJrh7nUepCBc golang.org/x/sys v0.0.0-20220319134239-a9b59b0215f8/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20220422013727-9388b58f7150/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20220520151302-bc2c85ada10a/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= -golang.org/x/sys v0.0.0-20220715151400-c0bba94af5f8/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20220722155257-8c9f86f7a55f/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.0.0-20220811171246-fbc7d0a398ab/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= golang.org/x/sys v0.1.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= diff --git a/pkg/performanceprofile/profilecreator/alert.go b/pkg/performanceprofile/profilecreator/alert.go new file mode 100644 index 0000000000..f0b11cd495 --- /dev/null +++ b/pkg/performanceprofile/profilecreator/alert.go @@ -0,0 +1,30 @@ +package profilecreator + +import ( + "log" + "os" +) + +var alertSink *log.Logger + +func init() { + // intentionally no flag for minimal output + alertSink = log.New(os.Stderr, "PPC: ", 0) +} + +// Alert raise status messages. `profilecreator` is meant to run interactively, +// so the default behavior is to bubble up the messages on the user console. +// It is however possible to reroute the messages to any log.Logger compliant backend +// see the function SetAlertSink. All Alerts are considered Info messages. +func Alert(format string, values ...any) { + alertSink.Printf(format, values...) +} + +// Call this function before any other else in the `profilecreator` package +func SetAlertSink(lh *log.Logger) { + alertSink = lh +} + +func GetAlertSink() *log.Logger { + return alertSink +} diff --git a/pkg/performanceprofile/profilecreator/cmd/info.go b/pkg/performanceprofile/profilecreator/cmd/info.go index 91a048429e..2b5d0caa6f 100644 --- a/pkg/performanceprofile/profilecreator/cmd/info.go +++ b/pkg/performanceprofile/profilecreator/cmd/info.go @@ -7,7 +7,6 @@ import ( "sort" "strings" - log "github.com/sirupsen/logrus" "github.com/spf13/cobra" ) @@ -95,13 +94,13 @@ func makeClusterInfoFromClusterData(cluster ClusterData) ClusterInfo { for _, handle := range nodeHandlers { topology, err := handle.SortedTopology() if err != nil { - log.Infof("%s(Topology discovery error: %v)", handle.Node.GetName(), err) + Alert("%s(Topology discovery error: %v)", handle.Node.GetName(), err) continue } htEnabled, err := handle.IsHyperthreadingEnabled() if err != nil { - log.Infof("%s(HT discovery error: %v)", handle.Node.GetName(), err) + Alert("%s(HT discovery error: %v)", handle.Node.GetName(), err) } nInfo := NodeInfo{ @@ -136,7 +135,7 @@ func showClusterInfo(cInfo ClusterInfo, infoOpts *infoOptions) error { fmt.Print(textInfo) return nil } - log.Infof("Cluster info:\n%s", textInfo) + Alert("Cluster info:\n%s", textInfo) return nil } diff --git a/pkg/performanceprofile/profilecreator/cmd/root.go b/pkg/performanceprofile/profilecreator/cmd/root.go index 3ab5831efe..037bfad7b8 100644 --- a/pkg/performanceprofile/profilecreator/cmd/root.go +++ b/pkg/performanceprofile/profilecreator/cmd/root.go @@ -22,7 +22,6 @@ import ( "os" "strings" - log "github.com/sirupsen/logrus" "github.com/spf13/cobra" "github.com/spf13/pflag" @@ -105,11 +104,10 @@ type ProfileData struct { // ClusterData collects the cluster wide information, each mcp points to a list of ghw node handlers type ClusterData map[string][]*profilecreator.GHWHandler +// shortcut +var Alert = profilecreator.Alert + func init() { - log.SetOutput(os.Stderr) - log.SetFormatter(&log.TextFormatter{ - DisableTimestamp: true, - }) utilruntime.Must(performancev2.AddToScheme(scheme)) } @@ -204,7 +202,7 @@ func makeNodesHandlers(mustGatherDirPath, poolName string, nodes []*corev1.Node) sb.WriteString(node.Name + " ") } // NodePoolName is alias of MCPName - log.Infof("Nodes names targeted by %s pool are: %s", poolName, sb.String()) + Alert("Nodes names targeted by %s pool are: %s", poolName, sb.String()) return handlers, nil } @@ -314,8 +312,8 @@ func makeProfileDataFrom(nodeHandler *profilecreator.GHWHandler, args *ProfileCr if err != nil { return nil, fmt.Errorf("failed to compute the reserved and isolated CPUs: %v", err) } - log.Infof("%d reserved CPUs allocated: %v ", reservedCPUs.Size(), reservedCPUs.String()) - log.Infof("%d isolated CPUs allocated: %v", isolatedCPUs.Size(), isolatedCPUs.String()) + Alert("%d reserved CPUs allocated: %v ", reservedCPUs.Size(), reservedCPUs.String()) + Alert("%d isolated CPUs allocated: %v", isolatedCPUs.Size(), isolatedCPUs.String()) kernelArgs := profilecreator.GetAdditionalKernelArgs(args.DisableHT) profileData := &ProfileData{ reservedCPUs: reservedCPUs.String(), diff --git a/pkg/performanceprofile/profilecreator/mcp.go b/pkg/performanceprofile/profilecreator/mcp.go index c09fa95ae9..7543a69bcc 100644 --- a/pkg/performanceprofile/profilecreator/mcp.go +++ b/pkg/performanceprofile/profilecreator/mcp.go @@ -4,8 +4,6 @@ import ( "fmt" "strings" - log "github.com/sirupsen/logrus" - corev1 "k8s.io/api/core/v1" metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" "k8s.io/apimachinery/pkg/labels" @@ -77,7 +75,7 @@ func GetNodesForPool(pool *mcfgv1.MachineConfigPool, clusterPools []*mcfgv1.Mach for _, n := range clusterNodes { p, err := getPrimaryPoolForNode(n, clusterPools) if err != nil { - log.Warningf("can't get pool for node %q: %v", n.Name, err) + Alert("can't get pool for node %q: %v", n.Name, err) continue } if p == nil { @@ -121,7 +119,7 @@ func getPoolsForNode(node *corev1.Node, clusterPools []*mcfgv1.MachineConfigPool if isWindows(node) { // This is not an error, is this a Windows Node and it won't be managed by MCO. We're explicitly logging // here at a high level to disambiguate this from other pools = nil scenario - log.Infof("Node %v is a windows node so won't be managed by MCO", node.Name) + Alert("Node %v is a windows node so won't be managed by MCO", node.Name) return nil, nil } diff --git a/pkg/performanceprofile/profilecreator/mustgather.go b/pkg/performanceprofile/profilecreator/mustgather.go index 29574f5aa1..b99ed60ee7 100644 --- a/pkg/performanceprofile/profilecreator/mustgather.go +++ b/pkg/performanceprofile/profilecreator/mustgather.go @@ -24,7 +24,6 @@ import ( "strings" machineconfigv1 "github.com/openshift/api/machineconfiguration/v1" - log "github.com/sirupsen/logrus" v1 "k8s.io/api/core/v1" k8syaml "k8s.io/apimachinery/pkg/util/yaml" ) @@ -68,7 +67,7 @@ func getMustGatherFullPathsWithFilter(mustGatherPath string, suffix string, filt return "", fmt.Errorf("no match for the specified must gather directory path: %s and suffix: %s", mustGatherPath, suffix) } if len(paths) > 1 { - log.Warnf("Multiple matches for the specified must gather directory path: %s and suffix: %s", mustGatherPath, suffix) + Alert("Multiple matches for the specified must gather directory path: %s and suffix: %s", mustGatherPath, suffix) return "", fmt.Errorf("Multiple matches for the specified must gather directory path: %s and suffix: %s.\n Expected only one performance-addon-operator-must-gather* directory, please check the must-gather tarball", mustGatherPath, suffix) } // returning one possible path diff --git a/pkg/performanceprofile/profilecreator/profilecreator.go b/pkg/performanceprofile/profilecreator/profilecreator.go index 6fdc8a87ac..a2fea1a1b4 100644 --- a/pkg/performanceprofile/profilecreator/profilecreator.go +++ b/pkg/performanceprofile/profilecreator/profilecreator.go @@ -25,7 +25,6 @@ import ( "github.com/jaypipes/ghw/pkg/cpu" "github.com/jaypipes/ghw/pkg/topology" - log "github.com/sirupsen/logrus" configv1 "github.com/openshift/api/config/v1" k8syaml "k8s.io/apimachinery/pkg/util/yaml" @@ -239,7 +238,7 @@ func getOfflinedCPUs(extCpuInfo *extendedCPUInfo, offlinedCPUCount int, disableH } if lpOfflined < offlinedCPUCount { - log.Warnf("could not offline enough logical processors (required:%d, offlined:%d)", offlinedCPUCount, lpOfflined) + Alert("could not offline enough logical processors (required:%d, offlined:%d)", offlinedCPUCount, lpOfflined) } return offlined.Result(), nil } @@ -248,7 +247,7 @@ func updateTopologyInfo(topoInfo *topology.Info, disableHTFlag bool, htEnabled b //currently HT is enabled on the system and the user wants to disable HT if htEnabled && disableHTFlag { - log.Infof("Updating Topology info because currently hyperthreading is enabled and the performance profile will disable it") + Alert("Updating Topology info because currently hyperthreading is enabled and the performance profile will disable it") return topologyHTDisabled(topoInfo), nil } return topoInfo, nil @@ -256,21 +255,21 @@ func updateTopologyInfo(topoInfo *topology.Info, disableHTFlag bool, htEnabled b func getReservedCPUs(topologyInfo *topology.Info, reservedCPUCount int, splitReservedCPUsAcrossNUMA bool, disableHTFlag bool, htEnabled bool) (cpuset.CPUSet, error) { if htEnabled && disableHTFlag { - log.Infof("Currently hyperthreading is enabled and the performance profile will disable it") + Alert("Currently hyperthreading is enabled and the performance profile will disable it") htEnabled = false } - log.Infof("NUMA cell(s): %d", len(topologyInfo.Nodes)) + Alert("NUMA cell(s): %d", len(topologyInfo.Nodes)) totalCPUs := 0 for id, node := range topologyInfo.Nodes { coreList := []int{} for _, core := range node.Cores { coreList = append(coreList, core.LogicalProcessors...) } - log.Infof("NUMA cell %d : %v", id, coreList) + Alert("NUMA cell %d : %v", id, coreList) totalCPUs += len(coreList) } - log.Infof("CPU(s): %d", totalCPUs) + Alert("CPU(s): %d", totalCPUs) if splitReservedCPUsAcrossNUMA { res, err := getCPUsSplitAcrossNUMA(reservedCPUCount, htEnabled, topologyInfo.Nodes) @@ -351,7 +350,7 @@ func getCPUsSplitAcrossNUMA(reservedCPUCount int, htEnabled bool, topologyInfoNo reservedPerNuma := reservedCPUCount / numaNodeNum remainder := reservedCPUCount % numaNodeNum if remainder != 0 { - log.Warnf("The reserved CPUs cannot be split equally across NUMA Nodes") + Alert("The reserved CPUs cannot be split equally across NUMA Nodes") } for numaID, node := range topologyInfoNodes { if remainder != 0 { @@ -467,7 +466,7 @@ func ensureSameTopology(topology1, topology2 *topology.Info, tols toleration.Set // as the NUMA cells have same logical processors' count and IDs and same threads' number, // core ID equality is treated as best effort. That is because when scheduling workloads, // we care about the logical processors ids and their location on the NUMAs. - log.Warnf("the CPU core ids in NUMA node %d differ: %d vs %d", node1.ID, core1.ID, cores2[j].ID) + Alert("the CPU core ids in NUMA node %d differ: %d vs %d", node1.ID, core1.ID, cores2[j].ID) tols[toleration.DifferentCoreIDs] = true } if core1.NumThreads != cores2[j].NumThreads { @@ -488,7 +487,7 @@ func GetAdditionalKernelArgs(disableHT bool) []string { kernelArgs = append(kernelArgs, noSMTKernelArg) } sort.Strings(kernelArgs) - log.Infof("Additional Kernel Args based on configuration: %v", kernelArgs) + Alert("Additional Kernel Args based on configuration: %v", kernelArgs) return kernelArgs } diff --git a/pkg/performanceprofile/profilecreator/profilecreator_test.go b/pkg/performanceprofile/profilecreator/profilecreator_test.go index 8bbc99d607..43491b09e6 100644 --- a/pkg/performanceprofile/profilecreator/profilecreator_test.go +++ b/pkg/performanceprofile/profilecreator/profilecreator_test.go @@ -12,12 +12,12 @@ import ( v1 "k8s.io/api/core/v1" "k8s.io/apimachinery/pkg/util/sets" + "k8s.io/klog/v2" "k8s.io/utils/cpuset" "github.com/jaypipes/ghw/pkg/cpu" "github.com/jaypipes/ghw/pkg/topology" mcfgv1 "github.com/openshift/api/machineconfiguration/v1" - log "github.com/sirupsen/logrus" "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/controller/performanceprofile/components" "github.com/openshift/cluster-node-tuning-operator/pkg/performanceprofile/profilecreator/toleration" @@ -601,16 +601,16 @@ var _ = Describe("Performance profile creator: test with a simple cpu architectu reserved, isolated, offlined, err := CalculateCPUSets(&sysInfo, reservedCPUCount, offlinedCPUCount, splitReservedCPUsAcrossNUMA, disableHT, highPowerConsumptionMode) Expect(err).ToNot(HaveOccurred()) - log.Infof("Input:") - log.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) - log.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) - log.Infof("\tdisable-ht:\t\t\t%v", disableHT) - log.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) - log.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) - log.Infof("Output:") - log.Infof("\treserved: %s", reserved.String()) - log.Infof("\tisolated: %s", isolated.String()) - log.Infof("\tofflined: %s", offlined.String()) + klog.Infof("Input:") + klog.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) + klog.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) + klog.Infof("\tdisable-ht:\t\t\t%v", disableHT) + klog.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) + klog.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) + klog.Infof("Output:") + klog.Infof("\treserved: %s", reserved.String()) + klog.Infof("\tisolated: %s", isolated.String()) + klog.Infof("\tofflined: %s", offlined.String()) Expect(offlined.Intersection(reserved).IsEmpty()).To(BeTrue()) Expect(offlined.Intersection(isolated).IsEmpty()).To(BeTrue()) @@ -647,16 +647,16 @@ var _ = Describe("Performance profile creator: test with a simple cpu architectu reserved, isolated, offlined, err := CalculateCPUSets(&sysInfo, reservedCPUCount, offlinedCPUCount, splitReservedCPUsAcrossNUMA, disableHT, highPowerConsumptionMode) Expect(err).ToNot(HaveOccurred()) - log.Infof("Input:") - log.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) - log.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) - log.Infof("\tdisable-ht:\t\t\t%v", disableHT) - log.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) - log.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) - log.Infof("Output:") - log.Infof("\treserved: %s", reserved.String()) - log.Infof("\tisolated: %s", isolated.String()) - log.Infof("\tofflined: %s", offlined.String()) + klog.Infof("Input:") + klog.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) + klog.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) + klog.Infof("\tdisable-ht:\t\t\t%v", disableHT) + klog.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) + klog.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) + klog.Infof("Output:") + klog.Infof("\treserved: %s", reserved.String()) + klog.Infof("\tisolated: %s", isolated.String()) + klog.Infof("\tofflined: %s", offlined.String()) Expect(offlined.Intersection(reserved).IsEmpty()).To(BeTrue()) Expect(offlined.Intersection(isolated).IsEmpty()).To(BeTrue()) @@ -688,16 +688,16 @@ var _ = Describe("Performance profile creator: test with a simple cpu architectu reserved, isolated, offlined, err := CalculateCPUSets(&sysInfo, reservedCPUCount, offlinedCPUCount, splitReservedCPUsAcrossNUMA, disableHT, highPowerConsumptionMode) Expect(err).ToNot(HaveOccurred()) - log.Infof("Input:") - log.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) - log.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) - log.Infof("\tdisable-ht:\t\t\t%v", disableHT) - log.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) - log.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) - log.Infof("Output:") - log.Infof("\treserved: %s", reserved.String()) - log.Infof("\tisolated: %s", isolated.String()) - log.Infof("\tofflined: %s", offlined.String()) + klog.Infof("Input:") + klog.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) + klog.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) + klog.Infof("\tdisable-ht:\t\t\t%v", disableHT) + klog.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) + klog.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) + klog.Infof("Output:") + klog.Infof("\treserved: %s", reserved.String()) + klog.Infof("\tisolated: %s", isolated.String()) + klog.Infof("\tofflined: %s", offlined.String()) Expect(offlined.Intersection(reserved).IsEmpty()).To(BeTrue()) Expect(offlined.Intersection(isolated).IsEmpty()).To(BeTrue()) @@ -729,16 +729,16 @@ var _ = Describe("Performance profile creator: test with a simple cpu architectu reserved, isolated, offlined, err := CalculateCPUSets(&sysInfo, reservedCPUCount, offlinedCPUCount, splitReservedCPUsAcrossNUMA, disableHT, highPowerConsumptionMode) Expect(err).ToNot(HaveOccurred()) - log.Infof("Input:") - log.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) - log.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) - log.Infof("\tdisable-ht:\t\t\t%v", disableHT) - log.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) - log.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) - log.Infof("Output:") - log.Infof("\treserved: %s", reserved.String()) - log.Infof("\tisolated: %s", isolated.String()) - log.Infof("\tofflined: %s", offlined.String()) + klog.Infof("Input:") + klog.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) + klog.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) + klog.Infof("\tdisable-ht:\t\t\t%v", disableHT) + klog.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) + klog.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) + klog.Infof("Output:") + klog.Infof("\treserved: %s", reserved.String()) + klog.Infof("\tisolated: %s", isolated.String()) + klog.Infof("\tofflined: %s", offlined.String()) Expect(offlined.Intersection(reserved).IsEmpty()).To(BeTrue()) Expect(offlined.Intersection(isolated).IsEmpty()).To(BeTrue()) @@ -770,16 +770,16 @@ var _ = Describe("Performance profile creator: test with a simple cpu architectu reserved, isolated, offlined, err := CalculateCPUSets(&sysInfo, reservedCPUCount, offlinedCPUCount, splitReservedCPUsAcrossNUMA, disableHT, highPowerConsumptionMode) Expect(err).ToNot(HaveOccurred()) - log.Infof("Input:") - log.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) - log.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) - log.Infof("\tdisable-ht:\t\t\t%v", disableHT) - log.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) - log.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) - log.Infof("Output:") - log.Infof("\treserved: %s", reserved.String()) - log.Infof("\tisolated: %s", isolated.String()) - log.Infof("\tofflined: %s", offlined.String()) + klog.Infof("Input:") + klog.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) + klog.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) + klog.Infof("\tdisable-ht:\t\t\t%v", disableHT) + klog.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) + klog.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) + klog.Infof("Output:") + klog.Infof("\treserved: %s", reserved.String()) + klog.Infof("\tisolated: %s", isolated.String()) + klog.Infof("\tofflined: %s", offlined.String()) Expect(offlined.Intersection(reserved).IsEmpty()).To(BeTrue()) Expect(offlined.Intersection(isolated).IsEmpty()).To(BeTrue()) @@ -811,16 +811,16 @@ var _ = Describe("Performance profile creator: test with a simple cpu architectu reserved, isolated, offlined, err := CalculateCPUSets(&sysInfo, reservedCPUCount, offlinedCPUCount, splitReservedCPUsAcrossNUMA, disableHT, highPowerConsumptionMode) Expect(err).ToNot(HaveOccurred()) - log.Infof("Input:") - log.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) - log.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) - log.Infof("\tdisable-ht:\t\t\t%v", disableHT) - log.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) - log.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) - log.Infof("Output:") - log.Infof("\treserved: %s", reserved.String()) - log.Infof("\tisolated: %s", isolated.String()) - log.Infof("\tofflined: %s", offlined.String()) + klog.Infof("Input:") + klog.Infof("\treserved-cpu-count:\t\t%d", reservedCPUCount) + klog.Infof("\tofflined-cpu-count:\t\t%d", offlinedCPUCount) + klog.Infof("\tdisable-ht:\t\t\t%v", disableHT) + klog.Infof("\tsplitReservedCPUsAcrossNUMA:\t%v", splitReservedCPUsAcrossNUMA) + klog.Infof("\thighConsumptionMode:\t\t%v", highPowerConsumptionMode) + klog.Infof("Output:") + klog.Infof("\treserved: %s", reserved.String()) + klog.Infof("\tisolated: %s", isolated.String()) + klog.Infof("\tofflined: %s", offlined.String()) Expect(offlined.Intersection(reserved).IsEmpty()).To(BeTrue()) Expect(offlined.Intersection(isolated).IsEmpty()).To(BeTrue()) @@ -1156,9 +1156,9 @@ var _ = Describe("PerformanceProfileCreator: Populating Reserved and Isolated CP Expect(offlinedCPUSet.Intersection(isolatedCPUSet).IsEmpty()).To(BeTrue()) Expect(offlinedCPUSet.Size()).To(Equal(offlinedCPUCount)) - log.Infof("offlined:%s", offlinedCPUSet.String()) - log.Infof("reserved:%s", reservedCPUSet.String()) - log.Infof("isolated:%s", isolatedCPUSet.String()) + klog.Infof("offlined:%s", offlinedCPUSet.String()) + klog.Infof("reserved:%s", reservedCPUSet.String()) + klog.Infof("isolated:%s", isolatedCPUSet.String()) totalCPUSet, err := GetTotalCPUSetFromGHW(handle, disableHT) Expect(err).ToNot(HaveOccurred()) @@ -1196,9 +1196,9 @@ var _ = Describe("PerformanceProfileCreator: Populating Reserved and Isolated CP Expect(offlinedCPUSet.Intersection(isolatedCPUSet).IsEmpty()).To(BeTrue()) Expect(offlinedCPUSet.Size()).To(Equal(offlinedCPUCount)) - log.Infof("offlined:%s", offlinedCPUSet.String()) - log.Infof("reserved:%s", reservedCPUSet.String()) - log.Infof("isolated:%s", isolatedCPUSet.String()) + klog.Infof("offlined:%s", offlinedCPUSet.String()) + klog.Infof("reserved:%s", reservedCPUSet.String()) + klog.Infof("isolated:%s", isolatedCPUSet.String()) totalCPUSet, err := GetTotalCPUSetFromGHW(handle, disableHT) Expect(err).ToNot(HaveOccurred()) @@ -1232,9 +1232,9 @@ var _ = Describe("PerformanceProfileCreator: Populating Reserved and Isolated CP Expect(offlinedCPUSet.Intersection(isolatedCPUSet).IsEmpty()).To(BeTrue()) Expect(offlinedCPUSet.Size()).To(Equal(offlinedCPUCount)) - log.Infof("offlined:%s", offlinedCPUSet.String()) - log.Infof("reserved:%s", reservedCPUSet.String()) - log.Infof("isolated:%s", isolatedCPUSet.String()) + klog.Infof("offlined:%s", offlinedCPUSet.String()) + klog.Infof("reserved:%s", reservedCPUSet.String()) + klog.Infof("isolated:%s", isolatedCPUSet.String()) totalCPUSet, err := GetTotalCPUSetFromGHW(handle, disableHT) Expect(err).ToNot(HaveOccurred()) @@ -1268,9 +1268,9 @@ var _ = Describe("PerformanceProfileCreator: Populating Reserved and Isolated CP Expect(offlinedCPUSet.Intersection(isolatedCPUSet).IsEmpty()).To(BeTrue()) Expect(offlinedCPUSet.Size()).To(Equal(offlinedCPUCount)) - log.Infof("offlined:%s", offlinedCPUSet.String()) - log.Infof("reserved:%s", reservedCPUSet.String()) - log.Infof("isolated:%s", isolatedCPUSet.String()) + klog.Infof("offlined:%s", offlinedCPUSet.String()) + klog.Infof("reserved:%s", reservedCPUSet.String()) + klog.Infof("isolated:%s", isolatedCPUSet.String()) totalCPUSet, err := GetTotalCPUSetFromGHW(handle, disableHT) Expect(err).ToNot(HaveOccurred()) @@ -1304,9 +1304,9 @@ var _ = Describe("PerformanceProfileCreator: Populating Reserved and Isolated CP Expect(offlinedCPUSet.Intersection(isolatedCPUSet).IsEmpty()).To(BeTrue()) Expect(offlinedCPUSet.Size()).To(Equal(offlinedCPUCount)) - log.Infof("offlined:%s", offlinedCPUSet.String()) - log.Infof("reserved:%s", reservedCPUSet.String()) - log.Infof("isolated:%s", isolatedCPUSet.String()) + klog.Infof("offlined:%s", offlinedCPUSet.String()) + klog.Infof("reserved:%s", reservedCPUSet.String()) + klog.Infof("isolated:%s", isolatedCPUSet.String()) totalCPUSet, err := GetTotalCPUSetFromGHW(handle, disableHT) Expect(err).ToNot(HaveOccurred()) @@ -1339,9 +1339,9 @@ var _ = Describe("PerformanceProfileCreator: Populating Reserved and Isolated CP Expect(offlinedCPUSet.Intersection(isolatedCPUSet).IsEmpty()).To(BeTrue()) Expect(offlinedCPUSet.Size()).To(Equal(offlinedCPUCount)) - log.Infof("offlined:%s", offlinedCPUSet.String()) - log.Infof("reserved:%s", reservedCPUSet.String()) - log.Infof("isolated:%s", isolatedCPUSet.String()) + klog.Infof("offlined:%s", offlinedCPUSet.String()) + klog.Infof("reserved:%s", reservedCPUSet.String()) + klog.Infof("isolated:%s", isolatedCPUSet.String()) totalCPUSet, err := GetTotalCPUSetFromGHW(handle, disableHT) Expect(err).ToNot(HaveOccurred()) @@ -1369,9 +1369,9 @@ var _ = Describe("PerformanceProfileCreator: Populating Reserved and Isolated CP Expect(offlinedCPUSet.Intersection(isolatedCPUSet).IsEmpty()).To(BeTrue()) Expect(offlinedCPUSet.Size()).To(Equal(offlinedCPUCount)) - log.Infof("offlined:%s", offlinedCPUSet.String()) - log.Infof("reserved:%s", reservedCPUSet.String()) - log.Infof("isolated:%s", isolatedCPUSet.String()) + klog.Infof("offlined:%s", offlinedCPUSet.String()) + klog.Infof("reserved:%s", reservedCPUSet.String()) + klog.Infof("isolated:%s", isolatedCPUSet.String()) totalCPUSet, err := GetTotalCPUSetFromGHW(handle, disableHT) Expect(err).ToNot(HaveOccurred()) diff --git a/vendor/github.com/sirupsen/logrus/.gitignore b/vendor/github.com/sirupsen/logrus/.gitignore deleted file mode 100644 index 1fb13abebe..0000000000 --- a/vendor/github.com/sirupsen/logrus/.gitignore +++ /dev/null @@ -1,4 +0,0 @@ -logrus -vendor - -.idea/ diff --git a/vendor/github.com/sirupsen/logrus/.golangci.yml b/vendor/github.com/sirupsen/logrus/.golangci.yml deleted file mode 100644 index 65dc285037..0000000000 --- a/vendor/github.com/sirupsen/logrus/.golangci.yml +++ /dev/null @@ -1,40 +0,0 @@ -run: - # do not run on test files yet - tests: false - -# all available settings of specific linters -linters-settings: - errcheck: - # report about not checking of errors in type assetions: `a := b.(MyStruct)`; - # default is false: such cases aren't reported by default. - check-type-assertions: false - - # report about assignment of errors to blank identifier: `num, _ := strconv.Atoi(numStr)`; - # default is false: such cases aren't reported by default. - check-blank: false - - lll: - line-length: 100 - tab-width: 4 - - prealloc: - simple: false - range-loops: false - for-loops: false - - whitespace: - multi-if: false # Enforces newlines (or comments) after every multi-line if statement - multi-func: false # Enforces newlines (or comments) after every multi-line function signature - -linters: - enable: - - megacheck - - govet - disable: - - maligned - - prealloc - disable-all: false - presets: - - bugs - - unused - fast: false diff --git a/vendor/github.com/sirupsen/logrus/.travis.yml b/vendor/github.com/sirupsen/logrus/.travis.yml deleted file mode 100644 index c1dbd5a3a3..0000000000 --- a/vendor/github.com/sirupsen/logrus/.travis.yml +++ /dev/null @@ -1,15 +0,0 @@ -language: go -go_import_path: github.com/sirupsen/logrus -git: - depth: 1 -env: - - GO111MODULE=on -go: 1.15.x -os: linux -install: - - ./travis/install.sh -script: - - cd ci - - go run mage.go -v -w ../ crossBuild - - go run mage.go -v -w ../ lint - - go run mage.go -v -w ../ test diff --git a/vendor/github.com/sirupsen/logrus/CHANGELOG.md b/vendor/github.com/sirupsen/logrus/CHANGELOG.md deleted file mode 100644 index 7567f61289..0000000000 --- a/vendor/github.com/sirupsen/logrus/CHANGELOG.md +++ /dev/null @@ -1,259 +0,0 @@ -# 1.8.1 -Code quality: - * move magefile in its own subdir/submodule to remove magefile dependency on logrus consumer - * improve timestamp format documentation - -Fixes: - * fix race condition on logger hooks - - -# 1.8.0 - -Correct versioning number replacing v1.7.1. - -# 1.7.1 - -Beware this release has introduced a new public API and its semver is therefore incorrect. - -Code quality: - * use go 1.15 in travis - * use magefile as task runner - -Fixes: - * small fixes about new go 1.13 error formatting system - * Fix for long time race condiction with mutating data hooks - -Features: - * build support for zos - -# 1.7.0 -Fixes: - * the dependency toward a windows terminal library has been removed - -Features: - * a new buffer pool management API has been added - * a set of `Fn()` functions have been added - -# 1.6.0 -Fixes: - * end of line cleanup - * revert the entry concurrency bug fix whic leads to deadlock under some circumstances - * update dependency on go-windows-terminal-sequences to fix a crash with go 1.14 - -Features: - * add an option to the `TextFormatter` to completely disable fields quoting - -# 1.5.0 -Code quality: - * add golangci linter run on travis - -Fixes: - * add mutex for hooks concurrent access on `Entry` data - * caller function field for go1.14 - * fix build issue for gopherjs target - -Feature: - * add an hooks/writer sub-package whose goal is to split output on different stream depending on the trace level - * add a `DisableHTMLEscape` option in the `JSONFormatter` - * add `ForceQuote` and `PadLevelText` options in the `TextFormatter` - -# 1.4.2 - * Fixes build break for plan9, nacl, solaris -# 1.4.1 -This new release introduces: - * Enhance TextFormatter to not print caller information when they are empty (#944) - * Remove dependency on golang.org/x/crypto (#932, #943) - -Fixes: - * Fix Entry.WithContext method to return a copy of the initial entry (#941) - -# 1.4.0 -This new release introduces: - * Add `DeferExitHandler`, similar to `RegisterExitHandler` but prepending the handler to the list of handlers (semantically like `defer`) (#848). - * Add `CallerPrettyfier` to `JSONFormatter` and `TextFormatter` (#909, #911) - * Add `Entry.WithContext()` and `Entry.Context`, to set a context on entries to be used e.g. in hooks (#919). - -Fixes: - * Fix wrong method calls `Logger.Print` and `Logger.Warningln` (#893). - * Update `Entry.Logf` to not do string formatting unless the log level is enabled (#903) - * Fix infinite recursion on unknown `Level.String()` (#907) - * Fix race condition in `getCaller` (#916). - - -# 1.3.0 -This new release introduces: - * Log, Logf, Logln functions for Logger and Entry that take a Level - -Fixes: - * Building prometheus node_exporter on AIX (#840) - * Race condition in TextFormatter (#468) - * Travis CI import path (#868) - * Remove coloured output on Windows (#862) - * Pointer to func as field in JSONFormatter (#870) - * Properly marshal Levels (#873) - -# 1.2.0 -This new release introduces: - * A new method `SetReportCaller` in the `Logger` to enable the file, line and calling function from which the trace has been issued - * A new trace level named `Trace` whose level is below `Debug` - * A configurable exit function to be called upon a Fatal trace - * The `Level` object now implements `encoding.TextUnmarshaler` interface - -# 1.1.1 -This is a bug fix release. - * fix the build break on Solaris - * don't drop a whole trace in JSONFormatter when a field param is a function pointer which can not be serialized - -# 1.1.0 -This new release introduces: - * several fixes: - * a fix for a race condition on entry formatting - * proper cleanup of previously used entries before putting them back in the pool - * the extra new line at the end of message in text formatter has been removed - * a new global public API to check if a level is activated: IsLevelEnabled - * the following methods have been added to the Logger object - * IsLevelEnabled - * SetFormatter - * SetOutput - * ReplaceHooks - * introduction of go module - * an indent configuration for the json formatter - * output colour support for windows - * the field sort function is now configurable for text formatter - * the CLICOLOR and CLICOLOR\_FORCE environment variable support in text formater - -# 1.0.6 - -This new release introduces: - * a new api WithTime which allows to easily force the time of the log entry - which is mostly useful for logger wrapper - * a fix reverting the immutability of the entry given as parameter to the hooks - a new configuration field of the json formatter in order to put all the fields - in a nested dictionnary - * a new SetOutput method in the Logger - * a new configuration of the textformatter to configure the name of the default keys - * a new configuration of the text formatter to disable the level truncation - -# 1.0.5 - -* Fix hooks race (#707) -* Fix panic deadlock (#695) - -# 1.0.4 - -* Fix race when adding hooks (#612) -* Fix terminal check in AppEngine (#635) - -# 1.0.3 - -* Replace example files with testable examples - -# 1.0.2 - -* bug: quote non-string values in text formatter (#583) -* Make (*Logger) SetLevel a public method - -# 1.0.1 - -* bug: fix escaping in text formatter (#575) - -# 1.0.0 - -* Officially changed name to lower-case -* bug: colors on Windows 10 (#541) -* bug: fix race in accessing level (#512) - -# 0.11.5 - -* feature: add writer and writerlevel to entry (#372) - -# 0.11.4 - -* bug: fix undefined variable on solaris (#493) - -# 0.11.3 - -* formatter: configure quoting of empty values (#484) -* formatter: configure quoting character (default is `"`) (#484) -* bug: fix not importing io correctly in non-linux environments (#481) - -# 0.11.2 - -* bug: fix windows terminal detection (#476) - -# 0.11.1 - -* bug: fix tty detection with custom out (#471) - -# 0.11.0 - -* performance: Use bufferpool to allocate (#370) -* terminal: terminal detection for app-engine (#343) -* feature: exit handler (#375) - -# 0.10.0 - -* feature: Add a test hook (#180) -* feature: `ParseLevel` is now case-insensitive (#326) -* feature: `FieldLogger` interface that generalizes `Logger` and `Entry` (#308) -* performance: avoid re-allocations on `WithFields` (#335) - -# 0.9.0 - -* logrus/text_formatter: don't emit empty msg -* logrus/hooks/airbrake: move out of main repository -* logrus/hooks/sentry: move out of main repository -* logrus/hooks/papertrail: move out of main repository -* logrus/hooks/bugsnag: move out of main repository -* logrus/core: run tests with `-race` -* logrus/core: detect TTY based on `stderr` -* logrus/core: support `WithError` on logger -* logrus/core: Solaris support - -# 0.8.7 - -* logrus/core: fix possible race (#216) -* logrus/doc: small typo fixes and doc improvements - - -# 0.8.6 - -* hooks/raven: allow passing an initialized client - -# 0.8.5 - -* logrus/core: revert #208 - -# 0.8.4 - -* formatter/text: fix data race (#218) - -# 0.8.3 - -* logrus/core: fix entry log level (#208) -* logrus/core: improve performance of text formatter by 40% -* logrus/core: expose `LevelHooks` type -* logrus/core: add support for DragonflyBSD and NetBSD -* formatter/text: print structs more verbosely - -# 0.8.2 - -* logrus: fix more Fatal family functions - -# 0.8.1 - -* logrus: fix not exiting on `Fatalf` and `Fatalln` - -# 0.8.0 - -* logrus: defaults to stderr instead of stdout -* hooks/sentry: add special field for `*http.Request` -* formatter/text: ignore Windows for colors - -# 0.7.3 - -* formatter/\*: allow configuration of timestamp layout - -# 0.7.2 - -* formatter/text: Add configuration option for time format (#158) diff --git a/vendor/github.com/sirupsen/logrus/LICENSE b/vendor/github.com/sirupsen/logrus/LICENSE deleted file mode 100644 index f090cb42f3..0000000000 --- a/vendor/github.com/sirupsen/logrus/LICENSE +++ /dev/null @@ -1,21 +0,0 @@ -The MIT License (MIT) - -Copyright (c) 2014 Simon Eskildsen - -Permission is hereby granted, free of charge, to any person obtaining a copy -of this software and associated documentation files (the "Software"), to deal -in the Software without restriction, including without limitation the rights -to use, copy, modify, merge, publish, distribute, sublicense, and/or sell -copies of the Software, and to permit persons to whom the Software is -furnished to do so, subject to the following conditions: - -The above copyright notice and this permission notice shall be included in -all copies or substantial portions of the Software. - -THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR -IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, -FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE -AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER -LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, -OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN -THE SOFTWARE. diff --git a/vendor/github.com/sirupsen/logrus/README.md b/vendor/github.com/sirupsen/logrus/README.md deleted file mode 100644 index d1d4a85fd7..0000000000 --- a/vendor/github.com/sirupsen/logrus/README.md +++ /dev/null @@ -1,515 +0,0 @@ -# Logrus :walrus: [![Build Status](https://github.com/sirupsen/logrus/workflows/CI/badge.svg)](https://github.com/sirupsen/logrus/actions?query=workflow%3ACI) [![Build Status](https://travis-ci.org/sirupsen/logrus.svg?branch=master)](https://travis-ci.org/sirupsen/logrus) [![Go Reference](https://pkg.go.dev/badge/github.com/sirupsen/logrus.svg)](https://pkg.go.dev/github.com/sirupsen/logrus) - -Logrus is a structured logger for Go (golang), completely API compatible with -the standard library logger. - -**Logrus is in maintenance-mode.** We will not be introducing new features. It's -simply too hard to do in a way that won't break many people's projects, which is -the last thing you want from your Logging library (again...). - -This does not mean Logrus is dead. Logrus will continue to be maintained for -security, (backwards compatible) bug fixes, and performance (where we are -limited by the interface). - -I believe Logrus' biggest contribution is to have played a part in today's -widespread use of structured logging in Golang. There doesn't seem to be a -reason to do a major, breaking iteration into Logrus V2, since the fantastic Go -community has built those independently. Many fantastic alternatives have sprung -up. Logrus would look like those, had it been re-designed with what we know -about structured logging in Go today. Check out, for example, -[Zerolog][zerolog], [Zap][zap], and [Apex][apex]. - -[zerolog]: https://github.com/rs/zerolog -[zap]: https://github.com/uber-go/zap -[apex]: https://github.com/apex/log - -**Seeing weird case-sensitive problems?** It's in the past been possible to -import Logrus as both upper- and lower-case. Due to the Go package environment, -this caused issues in the community and we needed a standard. Some environments -experienced problems with the upper-case variant, so the lower-case was decided. -Everything using `logrus` will need to use the lower-case: -`github.com/sirupsen/logrus`. Any package that isn't, should be changed. - -To fix Glide, see [these -comments](https://github.com/sirupsen/logrus/issues/553#issuecomment-306591437). -For an in-depth explanation of the casing issue, see [this -comment](https://github.com/sirupsen/logrus/issues/570#issuecomment-313933276). - -Nicely color-coded in development (when a TTY is attached, otherwise just -plain text): - -![Colored](http://i.imgur.com/PY7qMwd.png) - -With `log.SetFormatter(&log.JSONFormatter{})`, for easy parsing by logstash -or Splunk: - -```text -{"animal":"walrus","level":"info","msg":"A group of walrus emerges from the -ocean","size":10,"time":"2014-03-10 19:57:38.562264131 -0400 EDT"} - -{"level":"warning","msg":"The group's number increased tremendously!", -"number":122,"omg":true,"time":"2014-03-10 19:57:38.562471297 -0400 EDT"} - -{"animal":"walrus","level":"info","msg":"A giant walrus appears!", -"size":10,"time":"2014-03-10 19:57:38.562500591 -0400 EDT"} - -{"animal":"walrus","level":"info","msg":"Tremendously sized cow enters the ocean.", -"size":9,"time":"2014-03-10 19:57:38.562527896 -0400 EDT"} - -{"level":"fatal","msg":"The ice breaks!","number":100,"omg":true, -"time":"2014-03-10 19:57:38.562543128 -0400 EDT"} -``` - -With the default `log.SetFormatter(&log.TextFormatter{})` when a TTY is not -attached, the output is compatible with the -[logfmt](http://godoc.org/github.com/kr/logfmt) format: - -```text -time="2015-03-26T01:27:38-04:00" level=debug msg="Started observing beach" animal=walrus number=8 -time="2015-03-26T01:27:38-04:00" level=info msg="A group of walrus emerges from the ocean" animal=walrus size=10 -time="2015-03-26T01:27:38-04:00" level=warning msg="The group's number increased tremendously!" number=122 omg=true -time="2015-03-26T01:27:38-04:00" level=debug msg="Temperature changes" temperature=-4 -time="2015-03-26T01:27:38-04:00" level=panic msg="It's over 9000!" animal=orca size=9009 -time="2015-03-26T01:27:38-04:00" level=fatal msg="The ice breaks!" err=&{0x2082280c0 map[animal:orca size:9009] 2015-03-26 01:27:38.441574009 -0400 EDT panic It's over 9000!} number=100 omg=true -``` -To ensure this behaviour even if a TTY is attached, set your formatter as follows: - -```go - log.SetFormatter(&log.TextFormatter{ - DisableColors: true, - FullTimestamp: true, - }) -``` - -#### Logging Method Name - -If you wish to add the calling method as a field, instruct the logger via: -```go -log.SetReportCaller(true) -``` -This adds the caller as 'method' like so: - -```json -{"animal":"penguin","level":"fatal","method":"github.com/sirupsen/arcticcreatures.migrate","msg":"a penguin swims by", -"time":"2014-03-10 19:57:38.562543129 -0400 EDT"} -``` - -```text -time="2015-03-26T01:27:38-04:00" level=fatal method=github.com/sirupsen/arcticcreatures.migrate msg="a penguin swims by" animal=penguin -``` -Note that this does add measurable overhead - the cost will depend on the version of Go, but is -between 20 and 40% in recent tests with 1.6 and 1.7. You can validate this in your -environment via benchmarks: -``` -go test -bench=.*CallerTracing -``` - - -#### Case-sensitivity - -The organization's name was changed to lower-case--and this will not be changed -back. If you are getting import conflicts due to case sensitivity, please use -the lower-case import: `github.com/sirupsen/logrus`. - -#### Example - -The simplest way to use Logrus is simply the package-level exported logger: - -```go -package main - -import ( - log "github.com/sirupsen/logrus" -) - -func main() { - log.WithFields(log.Fields{ - "animal": "walrus", - }).Info("A walrus appears") -} -``` - -Note that it's completely api-compatible with the stdlib logger, so you can -replace your `log` imports everywhere with `log "github.com/sirupsen/logrus"` -and you'll now have the flexibility of Logrus. You can customize it all you -want: - -```go -package main - -import ( - "os" - log "github.com/sirupsen/logrus" -) - -func init() { - // Log as JSON instead of the default ASCII formatter. - log.SetFormatter(&log.JSONFormatter{}) - - // Output to stdout instead of the default stderr - // Can be any io.Writer, see below for File example - log.SetOutput(os.Stdout) - - // Only log the warning severity or above. - log.SetLevel(log.WarnLevel) -} - -func main() { - log.WithFields(log.Fields{ - "animal": "walrus", - "size": 10, - }).Info("A group of walrus emerges from the ocean") - - log.WithFields(log.Fields{ - "omg": true, - "number": 122, - }).Warn("The group's number increased tremendously!") - - log.WithFields(log.Fields{ - "omg": true, - "number": 100, - }).Fatal("The ice breaks!") - - // A common pattern is to re-use fields between logging statements by re-using - // the logrus.Entry returned from WithFields() - contextLogger := log.WithFields(log.Fields{ - "common": "this is a common field", - "other": "I also should be logged always", - }) - - contextLogger.Info("I'll be logged with common and other field") - contextLogger.Info("Me too") -} -``` - -For more advanced usage such as logging to multiple locations from the same -application, you can also create an instance of the `logrus` Logger: - -```go -package main - -import ( - "os" - "github.com/sirupsen/logrus" -) - -// Create a new instance of the logger. You can have any number of instances. -var log = logrus.New() - -func main() { - // The API for setting attributes is a little different than the package level - // exported logger. See Godoc. - log.Out = os.Stdout - - // You could set this to any `io.Writer` such as a file - // file, err := os.OpenFile("logrus.log", os.O_CREATE|os.O_WRONLY|os.O_APPEND, 0666) - // if err == nil { - // log.Out = file - // } else { - // log.Info("Failed to log to file, using default stderr") - // } - - log.WithFields(logrus.Fields{ - "animal": "walrus", - "size": 10, - }).Info("A group of walrus emerges from the ocean") -} -``` - -#### Fields - -Logrus encourages careful, structured logging through logging fields instead of -long, unparseable error messages. For example, instead of: `log.Fatalf("Failed -to send event %s to topic %s with key %d")`, you should log the much more -discoverable: - -```go -log.WithFields(log.Fields{ - "event": event, - "topic": topic, - "key": key, -}).Fatal("Failed to send event") -``` - -We've found this API forces you to think about logging in a way that produces -much more useful logging messages. We've been in countless situations where just -a single added field to a log statement that was already there would've saved us -hours. The `WithFields` call is optional. - -In general, with Logrus using any of the `printf`-family functions should be -seen as a hint you should add a field, however, you can still use the -`printf`-family functions with Logrus. - -#### Default Fields - -Often it's helpful to have fields _always_ attached to log statements in an -application or parts of one. For example, you may want to always log the -`request_id` and `user_ip` in the context of a request. Instead of writing -`log.WithFields(log.Fields{"request_id": request_id, "user_ip": user_ip})` on -every line, you can create a `logrus.Entry` to pass around instead: - -```go -requestLogger := log.WithFields(log.Fields{"request_id": request_id, "user_ip": user_ip}) -requestLogger.Info("something happened on that request") # will log request_id and user_ip -requestLogger.Warn("something not great happened") -``` - -#### Hooks - -You can add hooks for logging levels. For example to send errors to an exception -tracking service on `Error`, `Fatal` and `Panic`, info to StatsD or log to -multiple places simultaneously, e.g. syslog. - -Logrus comes with [built-in hooks](hooks/). Add those, or your custom hook, in -`init`: - -```go -import ( - log "github.com/sirupsen/logrus" - "gopkg.in/gemnasium/logrus-airbrake-hook.v2" // the package is named "airbrake" - logrus_syslog "github.com/sirupsen/logrus/hooks/syslog" - "log/syslog" -) - -func init() { - - // Use the Airbrake hook to report errors that have Error severity or above to - // an exception tracker. You can create custom hooks, see the Hooks section. - log.AddHook(airbrake.NewHook(123, "xyz", "production")) - - hook, err := logrus_syslog.NewSyslogHook("udp", "localhost:514", syslog.LOG_INFO, "") - if err != nil { - log.Error("Unable to connect to local syslog daemon") - } else { - log.AddHook(hook) - } -} -``` -Note: Syslog hook also support connecting to local syslog (Ex. "/dev/log" or "/var/run/syslog" or "/var/run/log"). For the detail, please check the [syslog hook README](hooks/syslog/README.md). - -A list of currently known service hooks can be found in this wiki [page](https://github.com/sirupsen/logrus/wiki/Hooks) - - -#### Level logging - -Logrus has seven logging levels: Trace, Debug, Info, Warning, Error, Fatal and Panic. - -```go -log.Trace("Something very low level.") -log.Debug("Useful debugging information.") -log.Info("Something noteworthy happened!") -log.Warn("You should probably take a look at this.") -log.Error("Something failed but I'm not quitting.") -// Calls os.Exit(1) after logging -log.Fatal("Bye.") -// Calls panic() after logging -log.Panic("I'm bailing.") -``` - -You can set the logging level on a `Logger`, then it will only log entries with -that severity or anything above it: - -```go -// Will log anything that is info or above (warn, error, fatal, panic). Default. -log.SetLevel(log.InfoLevel) -``` - -It may be useful to set `log.Level = logrus.DebugLevel` in a debug or verbose -environment if your application has that. - -Note: If you want different log levels for global (`log.SetLevel(...)`) and syslog logging, please check the [syslog hook README](hooks/syslog/README.md#different-log-levels-for-local-and-remote-logging). - -#### Entries - -Besides the fields added with `WithField` or `WithFields` some fields are -automatically added to all logging events: - -1. `time`. The timestamp when the entry was created. -2. `msg`. The logging message passed to `{Info,Warn,Error,Fatal,Panic}` after - the `AddFields` call. E.g. `Failed to send event.` -3. `level`. The logging level. E.g. `info`. - -#### Environments - -Logrus has no notion of environment. - -If you wish for hooks and formatters to only be used in specific environments, -you should handle that yourself. For example, if your application has a global -variable `Environment`, which is a string representation of the environment you -could do: - -```go -import ( - log "github.com/sirupsen/logrus" -) - -func init() { - // do something here to set environment depending on an environment variable - // or command-line flag - if Environment == "production" { - log.SetFormatter(&log.JSONFormatter{}) - } else { - // The TextFormatter is default, you don't actually have to do this. - log.SetFormatter(&log.TextFormatter{}) - } -} -``` - -This configuration is how `logrus` was intended to be used, but JSON in -production is mostly only useful if you do log aggregation with tools like -Splunk or Logstash. - -#### Formatters - -The built-in logging formatters are: - -* `logrus.TextFormatter`. Logs the event in colors if stdout is a tty, otherwise - without colors. - * *Note:* to force colored output when there is no TTY, set the `ForceColors` - field to `true`. To force no colored output even if there is a TTY set the - `DisableColors` field to `true`. For Windows, see - [github.com/mattn/go-colorable](https://github.com/mattn/go-colorable). - * When colors are enabled, levels are truncated to 4 characters by default. To disable - truncation set the `DisableLevelTruncation` field to `true`. - * When outputting to a TTY, it's often helpful to visually scan down a column where all the levels are the same width. Setting the `PadLevelText` field to `true` enables this behavior, by adding padding to the level text. - * All options are listed in the [generated docs](https://godoc.org/github.com/sirupsen/logrus#TextFormatter). -* `logrus.JSONFormatter`. Logs fields as JSON. - * All options are listed in the [generated docs](https://godoc.org/github.com/sirupsen/logrus#JSONFormatter). - -Third party logging formatters: - -* [`FluentdFormatter`](https://github.com/joonix/log). Formats entries that can be parsed by Kubernetes and Google Container Engine. -* [`GELF`](https://github.com/fabienm/go-logrus-formatters). Formats entries so they comply to Graylog's [GELF 1.1 specification](http://docs.graylog.org/en/2.4/pages/gelf.html). -* [`logstash`](https://github.com/bshuster-repo/logrus-logstash-hook). Logs fields as [Logstash](http://logstash.net) Events. -* [`prefixed`](https://github.com/x-cray/logrus-prefixed-formatter). Displays log entry source along with alternative layout. -* [`zalgo`](https://github.com/aybabtme/logzalgo). Invoking the Power of Zalgo. -* [`nested-logrus-formatter`](https://github.com/antonfisher/nested-logrus-formatter). Converts logrus fields to a nested structure. -* [`powerful-logrus-formatter`](https://github.com/zput/zxcTool). get fileName, log's line number and the latest function's name when print log; Sava log to files. -* [`caption-json-formatter`](https://github.com/nolleh/caption_json_formatter). logrus's message json formatter with human-readable caption added. - -You can define your formatter by implementing the `Formatter` interface, -requiring a `Format` method. `Format` takes an `*Entry`. `entry.Data` is a -`Fields` type (`map[string]interface{}`) with all your fields as well as the -default ones (see Entries section above): - -```go -type MyJSONFormatter struct { -} - -log.SetFormatter(new(MyJSONFormatter)) - -func (f *MyJSONFormatter) Format(entry *Entry) ([]byte, error) { - // Note this doesn't include Time, Level and Message which are available on - // the Entry. Consult `godoc` on information about those fields or read the - // source of the official loggers. - serialized, err := json.Marshal(entry.Data) - if err != nil { - return nil, fmt.Errorf("Failed to marshal fields to JSON, %w", err) - } - return append(serialized, '\n'), nil -} -``` - -#### Logger as an `io.Writer` - -Logrus can be transformed into an `io.Writer`. That writer is the end of an `io.Pipe` and it is your responsibility to close it. - -```go -w := logger.Writer() -defer w.Close() - -srv := http.Server{ - // create a stdlib log.Logger that writes to - // logrus.Logger. - ErrorLog: log.New(w, "", 0), -} -``` - -Each line written to that writer will be printed the usual way, using formatters -and hooks. The level for those entries is `info`. - -This means that we can override the standard library logger easily: - -```go -logger := logrus.New() -logger.Formatter = &logrus.JSONFormatter{} - -// Use logrus for standard log output -// Note that `log` here references stdlib's log -// Not logrus imported under the name `log`. -log.SetOutput(logger.Writer()) -``` - -#### Rotation - -Log rotation is not provided with Logrus. Log rotation should be done by an -external program (like `logrotate(8)`) that can compress and delete old log -entries. It should not be a feature of the application-level logger. - -#### Tools - -| Tool | Description | -| ---- | ----------- | -|[Logrus Mate](https://github.com/gogap/logrus_mate)|Logrus mate is a tool for Logrus to manage loggers, you can initial logger's level, hook and formatter by config file, the logger will be generated with different configs in different environments.| -|[Logrus Viper Helper](https://github.com/heirko/go-contrib/tree/master/logrusHelper)|An Helper around Logrus to wrap with spf13/Viper to load configuration with fangs! And to simplify Logrus configuration use some behavior of [Logrus Mate](https://github.com/gogap/logrus_mate). [sample](https://github.com/heirko/iris-contrib/blob/master/middleware/logrus-logger/example) | - -#### Testing - -Logrus has a built in facility for asserting the presence of log messages. This is implemented through the `test` hook and provides: - -* decorators for existing logger (`test.NewLocal` and `test.NewGlobal`) which basically just adds the `test` hook -* a test logger (`test.NewNullLogger`) that just records log messages (and does not output any): - -```go -import( - "github.com/sirupsen/logrus" - "github.com/sirupsen/logrus/hooks/test" - "github.com/stretchr/testify/assert" - "testing" -) - -func TestSomething(t*testing.T){ - logger, hook := test.NewNullLogger() - logger.Error("Helloerror") - - assert.Equal(t, 1, len(hook.Entries)) - assert.Equal(t, logrus.ErrorLevel, hook.LastEntry().Level) - assert.Equal(t, "Helloerror", hook.LastEntry().Message) - - hook.Reset() - assert.Nil(t, hook.LastEntry()) -} -``` - -#### Fatal handlers - -Logrus can register one or more functions that will be called when any `fatal` -level message is logged. The registered handlers will be executed before -logrus performs an `os.Exit(1)`. This behavior may be helpful if callers need -to gracefully shutdown. Unlike a `panic("Something went wrong...")` call which can be intercepted with a deferred `recover` a call to `os.Exit(1)` can not be intercepted. - -``` -... -handler := func() { - // gracefully shutdown something... -} -logrus.RegisterExitHandler(handler) -... -``` - -#### Thread safety - -By default, Logger is protected by a mutex for concurrent writes. The mutex is held when calling hooks and writing logs. -If you are sure such locking is not needed, you can call logger.SetNoLock() to disable the locking. - -Situation when locking is not needed includes: - -* You have no hooks registered, or hooks calling is already thread-safe. - -* Writing to logger.Out is already thread-safe, for example: - - 1) logger.Out is protected by locks. - - 2) logger.Out is an os.File handler opened with `O_APPEND` flag, and every write is smaller than 4k. (This allows multi-thread/multi-process writing) - - (Refer to http://www.notthewizard.com/2014/06/17/are-files-appends-really-atomic/) diff --git a/vendor/github.com/sirupsen/logrus/alt_exit.go b/vendor/github.com/sirupsen/logrus/alt_exit.go deleted file mode 100644 index 8fd189e1cc..0000000000 --- a/vendor/github.com/sirupsen/logrus/alt_exit.go +++ /dev/null @@ -1,76 +0,0 @@ -package logrus - -// The following code was sourced and modified from the -// https://github.com/tebeka/atexit package governed by the following license: -// -// Copyright (c) 2012 Miki Tebeka . -// -// Permission is hereby granted, free of charge, to any person obtaining a copy of -// this software and associated documentation files (the "Software"), to deal in -// the Software without restriction, including without limitation the rights to -// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of -// the Software, and to permit persons to whom the Software is furnished to do so, -// subject to the following conditions: -// -// The above copyright notice and this permission notice shall be included in all -// copies or substantial portions of the Software. -// -// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR -// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS -// FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR -// COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER -// IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN -// CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. - -import ( - "fmt" - "os" -) - -var handlers = []func(){} - -func runHandler(handler func()) { - defer func() { - if err := recover(); err != nil { - fmt.Fprintln(os.Stderr, "Error: Logrus exit handler error:", err) - } - }() - - handler() -} - -func runHandlers() { - for _, handler := range handlers { - runHandler(handler) - } -} - -// Exit runs all the Logrus atexit handlers and then terminates the program using os.Exit(code) -func Exit(code int) { - runHandlers() - os.Exit(code) -} - -// RegisterExitHandler appends a Logrus Exit handler to the list of handlers, -// call logrus.Exit to invoke all handlers. The handlers will also be invoked when -// any Fatal log entry is made. -// -// This method is useful when a caller wishes to use logrus to log a fatal -// message but also needs to gracefully shutdown. An example usecase could be -// closing database connections, or sending a alert that the application is -// closing. -func RegisterExitHandler(handler func()) { - handlers = append(handlers, handler) -} - -// DeferExitHandler prepends a Logrus Exit handler to the list of handlers, -// call logrus.Exit to invoke all handlers. The handlers will also be invoked when -// any Fatal log entry is made. -// -// This method is useful when a caller wishes to use logrus to log a fatal -// message but also needs to gracefully shutdown. An example usecase could be -// closing database connections, or sending a alert that the application is -// closing. -func DeferExitHandler(handler func()) { - handlers = append([]func(){handler}, handlers...) -} diff --git a/vendor/github.com/sirupsen/logrus/appveyor.yml b/vendor/github.com/sirupsen/logrus/appveyor.yml deleted file mode 100644 index df9d65c3a5..0000000000 --- a/vendor/github.com/sirupsen/logrus/appveyor.yml +++ /dev/null @@ -1,14 +0,0 @@ -version: "{build}" -platform: x64 -clone_folder: c:\gopath\src\github.com\sirupsen\logrus -environment: - GOPATH: c:\gopath -branches: - only: - - master -install: - - set PATH=%GOPATH%\bin;c:\go\bin;%PATH% - - go version -build_script: - - go get -t - - go test diff --git a/vendor/github.com/sirupsen/logrus/buffer_pool.go b/vendor/github.com/sirupsen/logrus/buffer_pool.go deleted file mode 100644 index c7787f77cb..0000000000 --- a/vendor/github.com/sirupsen/logrus/buffer_pool.go +++ /dev/null @@ -1,43 +0,0 @@ -package logrus - -import ( - "bytes" - "sync" -) - -var ( - bufferPool BufferPool -) - -type BufferPool interface { - Put(*bytes.Buffer) - Get() *bytes.Buffer -} - -type defaultPool struct { - pool *sync.Pool -} - -func (p *defaultPool) Put(buf *bytes.Buffer) { - p.pool.Put(buf) -} - -func (p *defaultPool) Get() *bytes.Buffer { - return p.pool.Get().(*bytes.Buffer) -} - -// SetBufferPool allows to replace the default logrus buffer pool -// to better meets the specific needs of an application. -func SetBufferPool(bp BufferPool) { - bufferPool = bp -} - -func init() { - SetBufferPool(&defaultPool{ - pool: &sync.Pool{ - New: func() interface{} { - return new(bytes.Buffer) - }, - }, - }) -} diff --git a/vendor/github.com/sirupsen/logrus/doc.go b/vendor/github.com/sirupsen/logrus/doc.go deleted file mode 100644 index da67aba06d..0000000000 --- a/vendor/github.com/sirupsen/logrus/doc.go +++ /dev/null @@ -1,26 +0,0 @@ -/* -Package logrus is a structured logger for Go, completely API compatible with the standard library logger. - - -The simplest way to use Logrus is simply the package-level exported logger: - - package main - - import ( - log "github.com/sirupsen/logrus" - ) - - func main() { - log.WithFields(log.Fields{ - "animal": "walrus", - "number": 1, - "size": 10, - }).Info("A walrus appears") - } - -Output: - time="2015-09-07T08:48:33Z" level=info msg="A walrus appears" animal=walrus number=1 size=10 - -For a full guide visit https://github.com/sirupsen/logrus -*/ -package logrus diff --git a/vendor/github.com/sirupsen/logrus/entry.go b/vendor/github.com/sirupsen/logrus/entry.go deleted file mode 100644 index 71cdbbc35d..0000000000 --- a/vendor/github.com/sirupsen/logrus/entry.go +++ /dev/null @@ -1,442 +0,0 @@ -package logrus - -import ( - "bytes" - "context" - "fmt" - "os" - "reflect" - "runtime" - "strings" - "sync" - "time" -) - -var ( - - // qualified package name, cached at first use - logrusPackage string - - // Positions in the call stack when tracing to report the calling method - minimumCallerDepth int - - // Used for caller information initialisation - callerInitOnce sync.Once -) - -const ( - maximumCallerDepth int = 25 - knownLogrusFrames int = 4 -) - -func init() { - // start at the bottom of the stack before the package-name cache is primed - minimumCallerDepth = 1 -} - -// Defines the key when adding errors using WithError. -var ErrorKey = "error" - -// An entry is the final or intermediate Logrus logging entry. It contains all -// the fields passed with WithField{,s}. It's finally logged when Trace, Debug, -// Info, Warn, Error, Fatal or Panic is called on it. These objects can be -// reused and passed around as much as you wish to avoid field duplication. -type Entry struct { - Logger *Logger - - // Contains all the fields set by the user. - Data Fields - - // Time at which the log entry was created - Time time.Time - - // Level the log entry was logged at: Trace, Debug, Info, Warn, Error, Fatal or Panic - // This field will be set on entry firing and the value will be equal to the one in Logger struct field. - Level Level - - // Calling method, with package name - Caller *runtime.Frame - - // Message passed to Trace, Debug, Info, Warn, Error, Fatal or Panic - Message string - - // When formatter is called in entry.log(), a Buffer may be set to entry - Buffer *bytes.Buffer - - // Contains the context set by the user. Useful for hook processing etc. - Context context.Context - - // err may contain a field formatting error - err string -} - -func NewEntry(logger *Logger) *Entry { - return &Entry{ - Logger: logger, - // Default is three fields, plus one optional. Give a little extra room. - Data: make(Fields, 6), - } -} - -func (entry *Entry) Dup() *Entry { - data := make(Fields, len(entry.Data)) - for k, v := range entry.Data { - data[k] = v - } - return &Entry{Logger: entry.Logger, Data: data, Time: entry.Time, Context: entry.Context, err: entry.err} -} - -// Returns the bytes representation of this entry from the formatter. -func (entry *Entry) Bytes() ([]byte, error) { - return entry.Logger.Formatter.Format(entry) -} - -// Returns the string representation from the reader and ultimately the -// formatter. -func (entry *Entry) String() (string, error) { - serialized, err := entry.Bytes() - if err != nil { - return "", err - } - str := string(serialized) - return str, nil -} - -// Add an error as single field (using the key defined in ErrorKey) to the Entry. -func (entry *Entry) WithError(err error) *Entry { - return entry.WithField(ErrorKey, err) -} - -// Add a context to the Entry. -func (entry *Entry) WithContext(ctx context.Context) *Entry { - dataCopy := make(Fields, len(entry.Data)) - for k, v := range entry.Data { - dataCopy[k] = v - } - return &Entry{Logger: entry.Logger, Data: dataCopy, Time: entry.Time, err: entry.err, Context: ctx} -} - -// Add a single field to the Entry. -func (entry *Entry) WithField(key string, value interface{}) *Entry { - return entry.WithFields(Fields{key: value}) -} - -// Add a map of fields to the Entry. -func (entry *Entry) WithFields(fields Fields) *Entry { - data := make(Fields, len(entry.Data)+len(fields)) - for k, v := range entry.Data { - data[k] = v - } - fieldErr := entry.err - for k, v := range fields { - isErrField := false - if t := reflect.TypeOf(v); t != nil { - switch { - case t.Kind() == reflect.Func, t.Kind() == reflect.Ptr && t.Elem().Kind() == reflect.Func: - isErrField = true - } - } - if isErrField { - tmp := fmt.Sprintf("can not add field %q", k) - if fieldErr != "" { - fieldErr = entry.err + ", " + tmp - } else { - fieldErr = tmp - } - } else { - data[k] = v - } - } - return &Entry{Logger: entry.Logger, Data: data, Time: entry.Time, err: fieldErr, Context: entry.Context} -} - -// Overrides the time of the Entry. -func (entry *Entry) WithTime(t time.Time) *Entry { - dataCopy := make(Fields, len(entry.Data)) - for k, v := range entry.Data { - dataCopy[k] = v - } - return &Entry{Logger: entry.Logger, Data: dataCopy, Time: t, err: entry.err, Context: entry.Context} -} - -// getPackageName reduces a fully qualified function name to the package name -// There really ought to be to be a better way... -func getPackageName(f string) string { - for { - lastPeriod := strings.LastIndex(f, ".") - lastSlash := strings.LastIndex(f, "/") - if lastPeriod > lastSlash { - f = f[:lastPeriod] - } else { - break - } - } - - return f -} - -// getCaller retrieves the name of the first non-logrus calling function -func getCaller() *runtime.Frame { - // cache this package's fully-qualified name - callerInitOnce.Do(func() { - pcs := make([]uintptr, maximumCallerDepth) - _ = runtime.Callers(0, pcs) - - // dynamic get the package name and the minimum caller depth - for i := 0; i < maximumCallerDepth; i++ { - funcName := runtime.FuncForPC(pcs[i]).Name() - if strings.Contains(funcName, "getCaller") { - logrusPackage = getPackageName(funcName) - break - } - } - - minimumCallerDepth = knownLogrusFrames - }) - - // Restrict the lookback frames to avoid runaway lookups - pcs := make([]uintptr, maximumCallerDepth) - depth := runtime.Callers(minimumCallerDepth, pcs) - frames := runtime.CallersFrames(pcs[:depth]) - - for f, again := frames.Next(); again; f, again = frames.Next() { - pkg := getPackageName(f.Function) - - // If the caller isn't part of this package, we're done - if pkg != logrusPackage { - return &f //nolint:scopelint - } - } - - // if we got here, we failed to find the caller's context - return nil -} - -func (entry Entry) HasCaller() (has bool) { - return entry.Logger != nil && - entry.Logger.ReportCaller && - entry.Caller != nil -} - -func (entry *Entry) log(level Level, msg string) { - var buffer *bytes.Buffer - - newEntry := entry.Dup() - - if newEntry.Time.IsZero() { - newEntry.Time = time.Now() - } - - newEntry.Level = level - newEntry.Message = msg - - newEntry.Logger.mu.Lock() - reportCaller := newEntry.Logger.ReportCaller - bufPool := newEntry.getBufferPool() - newEntry.Logger.mu.Unlock() - - if reportCaller { - newEntry.Caller = getCaller() - } - - newEntry.fireHooks() - buffer = bufPool.Get() - defer func() { - newEntry.Buffer = nil - buffer.Reset() - bufPool.Put(buffer) - }() - buffer.Reset() - newEntry.Buffer = buffer - - newEntry.write() - - newEntry.Buffer = nil - - // To avoid Entry#log() returning a value that only would make sense for - // panic() to use in Entry#Panic(), we avoid the allocation by checking - // directly here. - if level <= PanicLevel { - panic(newEntry) - } -} - -func (entry *Entry) getBufferPool() (pool BufferPool) { - if entry.Logger.BufferPool != nil { - return entry.Logger.BufferPool - } - return bufferPool -} - -func (entry *Entry) fireHooks() { - var tmpHooks LevelHooks - entry.Logger.mu.Lock() - tmpHooks = make(LevelHooks, len(entry.Logger.Hooks)) - for k, v := range entry.Logger.Hooks { - tmpHooks[k] = v - } - entry.Logger.mu.Unlock() - - err := tmpHooks.Fire(entry.Level, entry) - if err != nil { - fmt.Fprintf(os.Stderr, "Failed to fire hook: %v\n", err) - } -} - -func (entry *Entry) write() { - entry.Logger.mu.Lock() - defer entry.Logger.mu.Unlock() - serialized, err := entry.Logger.Formatter.Format(entry) - if err != nil { - fmt.Fprintf(os.Stderr, "Failed to obtain reader, %v\n", err) - return - } - if _, err := entry.Logger.Out.Write(serialized); err != nil { - fmt.Fprintf(os.Stderr, "Failed to write to log, %v\n", err) - } -} - -// Log will log a message at the level given as parameter. -// Warning: using Log at Panic or Fatal level will not respectively Panic nor Exit. -// For this behaviour Entry.Panic or Entry.Fatal should be used instead. -func (entry *Entry) Log(level Level, args ...interface{}) { - if entry.Logger.IsLevelEnabled(level) { - entry.log(level, fmt.Sprint(args...)) - } -} - -func (entry *Entry) Trace(args ...interface{}) { - entry.Log(TraceLevel, args...) -} - -func (entry *Entry) Debug(args ...interface{}) { - entry.Log(DebugLevel, args...) -} - -func (entry *Entry) Print(args ...interface{}) { - entry.Info(args...) -} - -func (entry *Entry) Info(args ...interface{}) { - entry.Log(InfoLevel, args...) -} - -func (entry *Entry) Warn(args ...interface{}) { - entry.Log(WarnLevel, args...) -} - -func (entry *Entry) Warning(args ...interface{}) { - entry.Warn(args...) -} - -func (entry *Entry) Error(args ...interface{}) { - entry.Log(ErrorLevel, args...) -} - -func (entry *Entry) Fatal(args ...interface{}) { - entry.Log(FatalLevel, args...) - entry.Logger.Exit(1) -} - -func (entry *Entry) Panic(args ...interface{}) { - entry.Log(PanicLevel, args...) -} - -// Entry Printf family functions - -func (entry *Entry) Logf(level Level, format string, args ...interface{}) { - if entry.Logger.IsLevelEnabled(level) { - entry.Log(level, fmt.Sprintf(format, args...)) - } -} - -func (entry *Entry) Tracef(format string, args ...interface{}) { - entry.Logf(TraceLevel, format, args...) -} - -func (entry *Entry) Debugf(format string, args ...interface{}) { - entry.Logf(DebugLevel, format, args...) -} - -func (entry *Entry) Infof(format string, args ...interface{}) { - entry.Logf(InfoLevel, format, args...) -} - -func (entry *Entry) Printf(format string, args ...interface{}) { - entry.Infof(format, args...) -} - -func (entry *Entry) Warnf(format string, args ...interface{}) { - entry.Logf(WarnLevel, format, args...) -} - -func (entry *Entry) Warningf(format string, args ...interface{}) { - entry.Warnf(format, args...) -} - -func (entry *Entry) Errorf(format string, args ...interface{}) { - entry.Logf(ErrorLevel, format, args...) -} - -func (entry *Entry) Fatalf(format string, args ...interface{}) { - entry.Logf(FatalLevel, format, args...) - entry.Logger.Exit(1) -} - -func (entry *Entry) Panicf(format string, args ...interface{}) { - entry.Logf(PanicLevel, format, args...) -} - -// Entry Println family functions - -func (entry *Entry) Logln(level Level, args ...interface{}) { - if entry.Logger.IsLevelEnabled(level) { - entry.Log(level, entry.sprintlnn(args...)) - } -} - -func (entry *Entry) Traceln(args ...interface{}) { - entry.Logln(TraceLevel, args...) -} - -func (entry *Entry) Debugln(args ...interface{}) { - entry.Logln(DebugLevel, args...) -} - -func (entry *Entry) Infoln(args ...interface{}) { - entry.Logln(InfoLevel, args...) -} - -func (entry *Entry) Println(args ...interface{}) { - entry.Infoln(args...) -} - -func (entry *Entry) Warnln(args ...interface{}) { - entry.Logln(WarnLevel, args...) -} - -func (entry *Entry) Warningln(args ...interface{}) { - entry.Warnln(args...) -} - -func (entry *Entry) Errorln(args ...interface{}) { - entry.Logln(ErrorLevel, args...) -} - -func (entry *Entry) Fatalln(args ...interface{}) { - entry.Logln(FatalLevel, args...) - entry.Logger.Exit(1) -} - -func (entry *Entry) Panicln(args ...interface{}) { - entry.Logln(PanicLevel, args...) -} - -// Sprintlnn => Sprint no newline. This is to get the behavior of how -// fmt.Sprintln where spaces are always added between operands, regardless of -// their type. Instead of vendoring the Sprintln implementation to spare a -// string allocation, we do the simplest thing. -func (entry *Entry) sprintlnn(args ...interface{}) string { - msg := fmt.Sprintln(args...) - return msg[:len(msg)-1] -} diff --git a/vendor/github.com/sirupsen/logrus/exported.go b/vendor/github.com/sirupsen/logrus/exported.go deleted file mode 100644 index 017c30ce67..0000000000 --- a/vendor/github.com/sirupsen/logrus/exported.go +++ /dev/null @@ -1,270 +0,0 @@ -package logrus - -import ( - "context" - "io" - "time" -) - -var ( - // std is the name of the standard logger in stdlib `log` - std = New() -) - -func StandardLogger() *Logger { - return std -} - -// SetOutput sets the standard logger output. -func SetOutput(out io.Writer) { - std.SetOutput(out) -} - -// SetFormatter sets the standard logger formatter. -func SetFormatter(formatter Formatter) { - std.SetFormatter(formatter) -} - -// SetReportCaller sets whether the standard logger will include the calling -// method as a field. -func SetReportCaller(include bool) { - std.SetReportCaller(include) -} - -// SetLevel sets the standard logger level. -func SetLevel(level Level) { - std.SetLevel(level) -} - -// GetLevel returns the standard logger level. -func GetLevel() Level { - return std.GetLevel() -} - -// IsLevelEnabled checks if the log level of the standard logger is greater than the level param -func IsLevelEnabled(level Level) bool { - return std.IsLevelEnabled(level) -} - -// AddHook adds a hook to the standard logger hooks. -func AddHook(hook Hook) { - std.AddHook(hook) -} - -// WithError creates an entry from the standard logger and adds an error to it, using the value defined in ErrorKey as key. -func WithError(err error) *Entry { - return std.WithField(ErrorKey, err) -} - -// WithContext creates an entry from the standard logger and adds a context to it. -func WithContext(ctx context.Context) *Entry { - return std.WithContext(ctx) -} - -// WithField creates an entry from the standard logger and adds a field to -// it. If you want multiple fields, use `WithFields`. -// -// Note that it doesn't log until you call Debug, Print, Info, Warn, Fatal -// or Panic on the Entry it returns. -func WithField(key string, value interface{}) *Entry { - return std.WithField(key, value) -} - -// WithFields creates an entry from the standard logger and adds multiple -// fields to it. This is simply a helper for `WithField`, invoking it -// once for each field. -// -// Note that it doesn't log until you call Debug, Print, Info, Warn, Fatal -// or Panic on the Entry it returns. -func WithFields(fields Fields) *Entry { - return std.WithFields(fields) -} - -// WithTime creates an entry from the standard logger and overrides the time of -// logs generated with it. -// -// Note that it doesn't log until you call Debug, Print, Info, Warn, Fatal -// or Panic on the Entry it returns. -func WithTime(t time.Time) *Entry { - return std.WithTime(t) -} - -// Trace logs a message at level Trace on the standard logger. -func Trace(args ...interface{}) { - std.Trace(args...) -} - -// Debug logs a message at level Debug on the standard logger. -func Debug(args ...interface{}) { - std.Debug(args...) -} - -// Print logs a message at level Info on the standard logger. -func Print(args ...interface{}) { - std.Print(args...) -} - -// Info logs a message at level Info on the standard logger. -func Info(args ...interface{}) { - std.Info(args...) -} - -// Warn logs a message at level Warn on the standard logger. -func Warn(args ...interface{}) { - std.Warn(args...) -} - -// Warning logs a message at level Warn on the standard logger. -func Warning(args ...interface{}) { - std.Warning(args...) -} - -// Error logs a message at level Error on the standard logger. -func Error(args ...interface{}) { - std.Error(args...) -} - -// Panic logs a message at level Panic on the standard logger. -func Panic(args ...interface{}) { - std.Panic(args...) -} - -// Fatal logs a message at level Fatal on the standard logger then the process will exit with status set to 1. -func Fatal(args ...interface{}) { - std.Fatal(args...) -} - -// TraceFn logs a message from a func at level Trace on the standard logger. -func TraceFn(fn LogFunction) { - std.TraceFn(fn) -} - -// DebugFn logs a message from a func at level Debug on the standard logger. -func DebugFn(fn LogFunction) { - std.DebugFn(fn) -} - -// PrintFn logs a message from a func at level Info on the standard logger. -func PrintFn(fn LogFunction) { - std.PrintFn(fn) -} - -// InfoFn logs a message from a func at level Info on the standard logger. -func InfoFn(fn LogFunction) { - std.InfoFn(fn) -} - -// WarnFn logs a message from a func at level Warn on the standard logger. -func WarnFn(fn LogFunction) { - std.WarnFn(fn) -} - -// WarningFn logs a message from a func at level Warn on the standard logger. -func WarningFn(fn LogFunction) { - std.WarningFn(fn) -} - -// ErrorFn logs a message from a func at level Error on the standard logger. -func ErrorFn(fn LogFunction) { - std.ErrorFn(fn) -} - -// PanicFn logs a message from a func at level Panic on the standard logger. -func PanicFn(fn LogFunction) { - std.PanicFn(fn) -} - -// FatalFn logs a message from a func at level Fatal on the standard logger then the process will exit with status set to 1. -func FatalFn(fn LogFunction) { - std.FatalFn(fn) -} - -// Tracef logs a message at level Trace on the standard logger. -func Tracef(format string, args ...interface{}) { - std.Tracef(format, args...) -} - -// Debugf logs a message at level Debug on the standard logger. -func Debugf(format string, args ...interface{}) { - std.Debugf(format, args...) -} - -// Printf logs a message at level Info on the standard logger. -func Printf(format string, args ...interface{}) { - std.Printf(format, args...) -} - -// Infof logs a message at level Info on the standard logger. -func Infof(format string, args ...interface{}) { - std.Infof(format, args...) -} - -// Warnf logs a message at level Warn on the standard logger. -func Warnf(format string, args ...interface{}) { - std.Warnf(format, args...) -} - -// Warningf logs a message at level Warn on the standard logger. -func Warningf(format string, args ...interface{}) { - std.Warningf(format, args...) -} - -// Errorf logs a message at level Error on the standard logger. -func Errorf(format string, args ...interface{}) { - std.Errorf(format, args...) -} - -// Panicf logs a message at level Panic on the standard logger. -func Panicf(format string, args ...interface{}) { - std.Panicf(format, args...) -} - -// Fatalf logs a message at level Fatal on the standard logger then the process will exit with status set to 1. -func Fatalf(format string, args ...interface{}) { - std.Fatalf(format, args...) -} - -// Traceln logs a message at level Trace on the standard logger. -func Traceln(args ...interface{}) { - std.Traceln(args...) -} - -// Debugln logs a message at level Debug on the standard logger. -func Debugln(args ...interface{}) { - std.Debugln(args...) -} - -// Println logs a message at level Info on the standard logger. -func Println(args ...interface{}) { - std.Println(args...) -} - -// Infoln logs a message at level Info on the standard logger. -func Infoln(args ...interface{}) { - std.Infoln(args...) -} - -// Warnln logs a message at level Warn on the standard logger. -func Warnln(args ...interface{}) { - std.Warnln(args...) -} - -// Warningln logs a message at level Warn on the standard logger. -func Warningln(args ...interface{}) { - std.Warningln(args...) -} - -// Errorln logs a message at level Error on the standard logger. -func Errorln(args ...interface{}) { - std.Errorln(args...) -} - -// Panicln logs a message at level Panic on the standard logger. -func Panicln(args ...interface{}) { - std.Panicln(args...) -} - -// Fatalln logs a message at level Fatal on the standard logger then the process will exit with status set to 1. -func Fatalln(args ...interface{}) { - std.Fatalln(args...) -} diff --git a/vendor/github.com/sirupsen/logrus/formatter.go b/vendor/github.com/sirupsen/logrus/formatter.go deleted file mode 100644 index 408883773e..0000000000 --- a/vendor/github.com/sirupsen/logrus/formatter.go +++ /dev/null @@ -1,78 +0,0 @@ -package logrus - -import "time" - -// Default key names for the default fields -const ( - defaultTimestampFormat = time.RFC3339 - FieldKeyMsg = "msg" - FieldKeyLevel = "level" - FieldKeyTime = "time" - FieldKeyLogrusError = "logrus_error" - FieldKeyFunc = "func" - FieldKeyFile = "file" -) - -// The Formatter interface is used to implement a custom Formatter. It takes an -// `Entry`. It exposes all the fields, including the default ones: -// -// * `entry.Data["msg"]`. The message passed from Info, Warn, Error .. -// * `entry.Data["time"]`. The timestamp. -// * `entry.Data["level"]. The level the entry was logged at. -// -// Any additional fields added with `WithField` or `WithFields` are also in -// `entry.Data`. Format is expected to return an array of bytes which are then -// logged to `logger.Out`. -type Formatter interface { - Format(*Entry) ([]byte, error) -} - -// This is to not silently overwrite `time`, `msg`, `func` and `level` fields when -// dumping it. If this code wasn't there doing: -// -// logrus.WithField("level", 1).Info("hello") -// -// Would just silently drop the user provided level. Instead with this code -// it'll logged as: -// -// {"level": "info", "fields.level": 1, "msg": "hello", "time": "..."} -// -// It's not exported because it's still using Data in an opinionated way. It's to -// avoid code duplication between the two default formatters. -func prefixFieldClashes(data Fields, fieldMap FieldMap, reportCaller bool) { - timeKey := fieldMap.resolve(FieldKeyTime) - if t, ok := data[timeKey]; ok { - data["fields."+timeKey] = t - delete(data, timeKey) - } - - msgKey := fieldMap.resolve(FieldKeyMsg) - if m, ok := data[msgKey]; ok { - data["fields."+msgKey] = m - delete(data, msgKey) - } - - levelKey := fieldMap.resolve(FieldKeyLevel) - if l, ok := data[levelKey]; ok { - data["fields."+levelKey] = l - delete(data, levelKey) - } - - logrusErrKey := fieldMap.resolve(FieldKeyLogrusError) - if l, ok := data[logrusErrKey]; ok { - data["fields."+logrusErrKey] = l - delete(data, logrusErrKey) - } - - // If reportCaller is not set, 'func' will not conflict. - if reportCaller { - funcKey := fieldMap.resolve(FieldKeyFunc) - if l, ok := data[funcKey]; ok { - data["fields."+funcKey] = l - } - fileKey := fieldMap.resolve(FieldKeyFile) - if l, ok := data[fileKey]; ok { - data["fields."+fileKey] = l - } - } -} diff --git a/vendor/github.com/sirupsen/logrus/hooks.go b/vendor/github.com/sirupsen/logrus/hooks.go deleted file mode 100644 index 3f151cdc39..0000000000 --- a/vendor/github.com/sirupsen/logrus/hooks.go +++ /dev/null @@ -1,34 +0,0 @@ -package logrus - -// A hook to be fired when logging on the logging levels returned from -// `Levels()` on your implementation of the interface. Note that this is not -// fired in a goroutine or a channel with workers, you should handle such -// functionality yourself if your call is non-blocking and you don't wish for -// the logging calls for levels returned from `Levels()` to block. -type Hook interface { - Levels() []Level - Fire(*Entry) error -} - -// Internal type for storing the hooks on a logger instance. -type LevelHooks map[Level][]Hook - -// Add a hook to an instance of logger. This is called with -// `log.Hooks.Add(new(MyHook))` where `MyHook` implements the `Hook` interface. -func (hooks LevelHooks) Add(hook Hook) { - for _, level := range hook.Levels() { - hooks[level] = append(hooks[level], hook) - } -} - -// Fire all the hooks for the passed level. Used by `entry.log` to fire -// appropriate hooks for a log entry. -func (hooks LevelHooks) Fire(level Level, entry *Entry) error { - for _, hook := range hooks[level] { - if err := hook.Fire(entry); err != nil { - return err - } - } - - return nil -} diff --git a/vendor/github.com/sirupsen/logrus/json_formatter.go b/vendor/github.com/sirupsen/logrus/json_formatter.go deleted file mode 100644 index c96dc5636b..0000000000 --- a/vendor/github.com/sirupsen/logrus/json_formatter.go +++ /dev/null @@ -1,128 +0,0 @@ -package logrus - -import ( - "bytes" - "encoding/json" - "fmt" - "runtime" -) - -type fieldKey string - -// FieldMap allows customization of the key names for default fields. -type FieldMap map[fieldKey]string - -func (f FieldMap) resolve(key fieldKey) string { - if k, ok := f[key]; ok { - return k - } - - return string(key) -} - -// JSONFormatter formats logs into parsable json -type JSONFormatter struct { - // TimestampFormat sets the format used for marshaling timestamps. - // The format to use is the same than for time.Format or time.Parse from the standard - // library. - // The standard Library already provides a set of predefined format. - TimestampFormat string - - // DisableTimestamp allows disabling automatic timestamps in output - DisableTimestamp bool - - // DisableHTMLEscape allows disabling html escaping in output - DisableHTMLEscape bool - - // DataKey allows users to put all the log entry parameters into a nested dictionary at a given key. - DataKey string - - // FieldMap allows users to customize the names of keys for default fields. - // As an example: - // formatter := &JSONFormatter{ - // FieldMap: FieldMap{ - // FieldKeyTime: "@timestamp", - // FieldKeyLevel: "@level", - // FieldKeyMsg: "@message", - // FieldKeyFunc: "@caller", - // }, - // } - FieldMap FieldMap - - // CallerPrettyfier can be set by the user to modify the content - // of the function and file keys in the json data when ReportCaller is - // activated. If any of the returned value is the empty string the - // corresponding key will be removed from json fields. - CallerPrettyfier func(*runtime.Frame) (function string, file string) - - // PrettyPrint will indent all json logs - PrettyPrint bool -} - -// Format renders a single log entry -func (f *JSONFormatter) Format(entry *Entry) ([]byte, error) { - data := make(Fields, len(entry.Data)+4) - for k, v := range entry.Data { - switch v := v.(type) { - case error: - // Otherwise errors are ignored by `encoding/json` - // https://github.com/sirupsen/logrus/issues/137 - data[k] = v.Error() - default: - data[k] = v - } - } - - if f.DataKey != "" { - newData := make(Fields, 4) - newData[f.DataKey] = data - data = newData - } - - prefixFieldClashes(data, f.FieldMap, entry.HasCaller()) - - timestampFormat := f.TimestampFormat - if timestampFormat == "" { - timestampFormat = defaultTimestampFormat - } - - if entry.err != "" { - data[f.FieldMap.resolve(FieldKeyLogrusError)] = entry.err - } - if !f.DisableTimestamp { - data[f.FieldMap.resolve(FieldKeyTime)] = entry.Time.Format(timestampFormat) - } - data[f.FieldMap.resolve(FieldKeyMsg)] = entry.Message - data[f.FieldMap.resolve(FieldKeyLevel)] = entry.Level.String() - if entry.HasCaller() { - funcVal := entry.Caller.Function - fileVal := fmt.Sprintf("%s:%d", entry.Caller.File, entry.Caller.Line) - if f.CallerPrettyfier != nil { - funcVal, fileVal = f.CallerPrettyfier(entry.Caller) - } - if funcVal != "" { - data[f.FieldMap.resolve(FieldKeyFunc)] = funcVal - } - if fileVal != "" { - data[f.FieldMap.resolve(FieldKeyFile)] = fileVal - } - } - - var b *bytes.Buffer - if entry.Buffer != nil { - b = entry.Buffer - } else { - b = &bytes.Buffer{} - } - - encoder := json.NewEncoder(b) - encoder.SetEscapeHTML(!f.DisableHTMLEscape) - if f.PrettyPrint { - encoder.SetIndent("", " ") - } - if err := encoder.Encode(data); err != nil { - return nil, fmt.Errorf("failed to marshal fields to JSON, %w", err) - } - - return b.Bytes(), nil -} diff --git a/vendor/github.com/sirupsen/logrus/logger.go b/vendor/github.com/sirupsen/logrus/logger.go deleted file mode 100644 index 5ff0aef6d3..0000000000 --- a/vendor/github.com/sirupsen/logrus/logger.go +++ /dev/null @@ -1,417 +0,0 @@ -package logrus - -import ( - "context" - "io" - "os" - "sync" - "sync/atomic" - "time" -) - -// LogFunction For big messages, it can be more efficient to pass a function -// and only call it if the log level is actually enables rather than -// generating the log message and then checking if the level is enabled -type LogFunction func() []interface{} - -type Logger struct { - // The logs are `io.Copy`'d to this in a mutex. It's common to set this to a - // file, or leave it default which is `os.Stderr`. You can also set this to - // something more adventurous, such as logging to Kafka. - Out io.Writer - // Hooks for the logger instance. These allow firing events based on logging - // levels and log entries. For example, to send errors to an error tracking - // service, log to StatsD or dump the core on fatal errors. - Hooks LevelHooks - // All log entries pass through the formatter before logged to Out. The - // included formatters are `TextFormatter` and `JSONFormatter` for which - // TextFormatter is the default. In development (when a TTY is attached) it - // logs with colors, but to a file it wouldn't. You can easily implement your - // own that implements the `Formatter` interface, see the `README` or included - // formatters for examples. - Formatter Formatter - - // Flag for whether to log caller info (off by default) - ReportCaller bool - - // The logging level the logger should log at. This is typically (and defaults - // to) `logrus.Info`, which allows Info(), Warn(), Error() and Fatal() to be - // logged. - Level Level - // Used to sync writing to the log. Locking is enabled by Default - mu MutexWrap - // Reusable empty entry - entryPool sync.Pool - // Function to exit the application, defaults to `os.Exit()` - ExitFunc exitFunc - // The buffer pool used to format the log. If it is nil, the default global - // buffer pool will be used. - BufferPool BufferPool -} - -type exitFunc func(int) - -type MutexWrap struct { - lock sync.Mutex - disabled bool -} - -func (mw *MutexWrap) Lock() { - if !mw.disabled { - mw.lock.Lock() - } -} - -func (mw *MutexWrap) Unlock() { - if !mw.disabled { - mw.lock.Unlock() - } -} - -func (mw *MutexWrap) Disable() { - mw.disabled = true -} - -// Creates a new logger. Configuration should be set by changing `Formatter`, -// `Out` and `Hooks` directly on the default logger instance. You can also just -// instantiate your own: -// -// var log = &logrus.Logger{ -// Out: os.Stderr, -// Formatter: new(logrus.TextFormatter), -// Hooks: make(logrus.LevelHooks), -// Level: logrus.DebugLevel, -// } -// -// It's recommended to make this a global instance called `log`. -func New() *Logger { - return &Logger{ - Out: os.Stderr, - Formatter: new(TextFormatter), - Hooks: make(LevelHooks), - Level: InfoLevel, - ExitFunc: os.Exit, - ReportCaller: false, - } -} - -func (logger *Logger) newEntry() *Entry { - entry, ok := logger.entryPool.Get().(*Entry) - if ok { - return entry - } - return NewEntry(logger) -} - -func (logger *Logger) releaseEntry(entry *Entry) { - entry.Data = map[string]interface{}{} - logger.entryPool.Put(entry) -} - -// WithField allocates a new entry and adds a field to it. -// Debug, Print, Info, Warn, Error, Fatal or Panic must be then applied to -// this new returned entry. -// If you want multiple fields, use `WithFields`. -func (logger *Logger) WithField(key string, value interface{}) *Entry { - entry := logger.newEntry() - defer logger.releaseEntry(entry) - return entry.WithField(key, value) -} - -// Adds a struct of fields to the log entry. All it does is call `WithField` for -// each `Field`. -func (logger *Logger) WithFields(fields Fields) *Entry { - entry := logger.newEntry() - defer logger.releaseEntry(entry) - return entry.WithFields(fields) -} - -// Add an error as single field to the log entry. All it does is call -// `WithError` for the given `error`. -func (logger *Logger) WithError(err error) *Entry { - entry := logger.newEntry() - defer logger.releaseEntry(entry) - return entry.WithError(err) -} - -// Add a context to the log entry. -func (logger *Logger) WithContext(ctx context.Context) *Entry { - entry := logger.newEntry() - defer logger.releaseEntry(entry) - return entry.WithContext(ctx) -} - -// Overrides the time of the log entry. -func (logger *Logger) WithTime(t time.Time) *Entry { - entry := logger.newEntry() - defer logger.releaseEntry(entry) - return entry.WithTime(t) -} - -func (logger *Logger) Logf(level Level, format string, args ...interface{}) { - if logger.IsLevelEnabled(level) { - entry := logger.newEntry() - entry.Logf(level, format, args...) - logger.releaseEntry(entry) - } -} - -func (logger *Logger) Tracef(format string, args ...interface{}) { - logger.Logf(TraceLevel, format, args...) -} - -func (logger *Logger) Debugf(format string, args ...interface{}) { - logger.Logf(DebugLevel, format, args...) -} - -func (logger *Logger) Infof(format string, args ...interface{}) { - logger.Logf(InfoLevel, format, args...) -} - -func (logger *Logger) Printf(format string, args ...interface{}) { - entry := logger.newEntry() - entry.Printf(format, args...) - logger.releaseEntry(entry) -} - -func (logger *Logger) Warnf(format string, args ...interface{}) { - logger.Logf(WarnLevel, format, args...) -} - -func (logger *Logger) Warningf(format string, args ...interface{}) { - logger.Warnf(format, args...) -} - -func (logger *Logger) Errorf(format string, args ...interface{}) { - logger.Logf(ErrorLevel, format, args...) -} - -func (logger *Logger) Fatalf(format string, args ...interface{}) { - logger.Logf(FatalLevel, format, args...) - logger.Exit(1) -} - -func (logger *Logger) Panicf(format string, args ...interface{}) { - logger.Logf(PanicLevel, format, args...) -} - -// Log will log a message at the level given as parameter. -// Warning: using Log at Panic or Fatal level will not respectively Panic nor Exit. -// For this behaviour Logger.Panic or Logger.Fatal should be used instead. -func (logger *Logger) Log(level Level, args ...interface{}) { - if logger.IsLevelEnabled(level) { - entry := logger.newEntry() - entry.Log(level, args...) - logger.releaseEntry(entry) - } -} - -func (logger *Logger) LogFn(level Level, fn LogFunction) { - if logger.IsLevelEnabled(level) { - entry := logger.newEntry() - entry.Log(level, fn()...) - logger.releaseEntry(entry) - } -} - -func (logger *Logger) Trace(args ...interface{}) { - logger.Log(TraceLevel, args...) -} - -func (logger *Logger) Debug(args ...interface{}) { - logger.Log(DebugLevel, args...) -} - -func (logger *Logger) Info(args ...interface{}) { - logger.Log(InfoLevel, args...) -} - -func (logger *Logger) Print(args ...interface{}) { - entry := logger.newEntry() - entry.Print(args...) - logger.releaseEntry(entry) -} - -func (logger *Logger) Warn(args ...interface{}) { - logger.Log(WarnLevel, args...) -} - -func (logger *Logger) Warning(args ...interface{}) { - logger.Warn(args...) -} - -func (logger *Logger) Error(args ...interface{}) { - logger.Log(ErrorLevel, args...) -} - -func (logger *Logger) Fatal(args ...interface{}) { - logger.Log(FatalLevel, args...) - logger.Exit(1) -} - -func (logger *Logger) Panic(args ...interface{}) { - logger.Log(PanicLevel, args...) -} - -func (logger *Logger) TraceFn(fn LogFunction) { - logger.LogFn(TraceLevel, fn) -} - -func (logger *Logger) DebugFn(fn LogFunction) { - logger.LogFn(DebugLevel, fn) -} - -func (logger *Logger) InfoFn(fn LogFunction) { - logger.LogFn(InfoLevel, fn) -} - -func (logger *Logger) PrintFn(fn LogFunction) { - entry := logger.newEntry() - entry.Print(fn()...) - logger.releaseEntry(entry) -} - -func (logger *Logger) WarnFn(fn LogFunction) { - logger.LogFn(WarnLevel, fn) -} - -func (logger *Logger) WarningFn(fn LogFunction) { - logger.WarnFn(fn) -} - -func (logger *Logger) ErrorFn(fn LogFunction) { - logger.LogFn(ErrorLevel, fn) -} - -func (logger *Logger) FatalFn(fn LogFunction) { - logger.LogFn(FatalLevel, fn) - logger.Exit(1) -} - -func (logger *Logger) PanicFn(fn LogFunction) { - logger.LogFn(PanicLevel, fn) -} - -func (logger *Logger) Logln(level Level, args ...interface{}) { - if logger.IsLevelEnabled(level) { - entry := logger.newEntry() - entry.Logln(level, args...) - logger.releaseEntry(entry) - } -} - -func (logger *Logger) Traceln(args ...interface{}) { - logger.Logln(TraceLevel, args...) -} - -func (logger *Logger) Debugln(args ...interface{}) { - logger.Logln(DebugLevel, args...) -} - -func (logger *Logger) Infoln(args ...interface{}) { - logger.Logln(InfoLevel, args...) -} - -func (logger *Logger) Println(args ...interface{}) { - entry := logger.newEntry() - entry.Println(args...) - logger.releaseEntry(entry) -} - -func (logger *Logger) Warnln(args ...interface{}) { - logger.Logln(WarnLevel, args...) -} - -func (logger *Logger) Warningln(args ...interface{}) { - logger.Warnln(args...) -} - -func (logger *Logger) Errorln(args ...interface{}) { - logger.Logln(ErrorLevel, args...) -} - -func (logger *Logger) Fatalln(args ...interface{}) { - logger.Logln(FatalLevel, args...) - logger.Exit(1) -} - -func (logger *Logger) Panicln(args ...interface{}) { - logger.Logln(PanicLevel, args...) -} - -func (logger *Logger) Exit(code int) { - runHandlers() - if logger.ExitFunc == nil { - logger.ExitFunc = os.Exit - } - logger.ExitFunc(code) -} - -//When file is opened with appending mode, it's safe to -//write concurrently to a file (within 4k message on Linux). -//In these cases user can choose to disable the lock. -func (logger *Logger) SetNoLock() { - logger.mu.Disable() -} - -func (logger *Logger) level() Level { - return Level(atomic.LoadUint32((*uint32)(&logger.Level))) -} - -// SetLevel sets the logger level. -func (logger *Logger) SetLevel(level Level) { - atomic.StoreUint32((*uint32)(&logger.Level), uint32(level)) -} - -// GetLevel returns the logger level. -func (logger *Logger) GetLevel() Level { - return logger.level() -} - -// AddHook adds a hook to the logger hooks. -func (logger *Logger) AddHook(hook Hook) { - logger.mu.Lock() - defer logger.mu.Unlock() - logger.Hooks.Add(hook) -} - -// IsLevelEnabled checks if the log level of the logger is greater than the level param -func (logger *Logger) IsLevelEnabled(level Level) bool { - return logger.level() >= level -} - -// SetFormatter sets the logger formatter. -func (logger *Logger) SetFormatter(formatter Formatter) { - logger.mu.Lock() - defer logger.mu.Unlock() - logger.Formatter = formatter -} - -// SetOutput sets the logger output. -func (logger *Logger) SetOutput(output io.Writer) { - logger.mu.Lock() - defer logger.mu.Unlock() - logger.Out = output -} - -func (logger *Logger) SetReportCaller(reportCaller bool) { - logger.mu.Lock() - defer logger.mu.Unlock() - logger.ReportCaller = reportCaller -} - -// ReplaceHooks replaces the logger hooks and returns the old ones -func (logger *Logger) ReplaceHooks(hooks LevelHooks) LevelHooks { - logger.mu.Lock() - oldHooks := logger.Hooks - logger.Hooks = hooks - logger.mu.Unlock() - return oldHooks -} - -// SetBufferPool sets the logger buffer pool. -func (logger *Logger) SetBufferPool(pool BufferPool) { - logger.mu.Lock() - defer logger.mu.Unlock() - logger.BufferPool = pool -} diff --git a/vendor/github.com/sirupsen/logrus/logrus.go b/vendor/github.com/sirupsen/logrus/logrus.go deleted file mode 100644 index 2f16224cb9..0000000000 --- a/vendor/github.com/sirupsen/logrus/logrus.go +++ /dev/null @@ -1,186 +0,0 @@ -package logrus - -import ( - "fmt" - "log" - "strings" -) - -// Fields type, used to pass to `WithFields`. -type Fields map[string]interface{} - -// Level type -type Level uint32 - -// Convert the Level to a string. E.g. PanicLevel becomes "panic". -func (level Level) String() string { - if b, err := level.MarshalText(); err == nil { - return string(b) - } else { - return "unknown" - } -} - -// ParseLevel takes a string level and returns the Logrus log level constant. -func ParseLevel(lvl string) (Level, error) { - switch strings.ToLower(lvl) { - case "panic": - return PanicLevel, nil - case "fatal": - return FatalLevel, nil - case "error": - return ErrorLevel, nil - case "warn", "warning": - return WarnLevel, nil - case "info": - return InfoLevel, nil - case "debug": - return DebugLevel, nil - case "trace": - return TraceLevel, nil - } - - var l Level - return l, fmt.Errorf("not a valid logrus Level: %q", lvl) -} - -// UnmarshalText implements encoding.TextUnmarshaler. -func (level *Level) UnmarshalText(text []byte) error { - l, err := ParseLevel(string(text)) - if err != nil { - return err - } - - *level = l - - return nil -} - -func (level Level) MarshalText() ([]byte, error) { - switch level { - case TraceLevel: - return []byte("trace"), nil - case DebugLevel: - return []byte("debug"), nil - case InfoLevel: - return []byte("info"), nil - case WarnLevel: - return []byte("warning"), nil - case ErrorLevel: - return []byte("error"), nil - case FatalLevel: - return []byte("fatal"), nil - case PanicLevel: - return []byte("panic"), nil - } - - return nil, fmt.Errorf("not a valid logrus level %d", level) -} - -// A constant exposing all logging levels -var AllLevels = []Level{ - PanicLevel, - FatalLevel, - ErrorLevel, - WarnLevel, - InfoLevel, - DebugLevel, - TraceLevel, -} - -// These are the different logging levels. You can set the logging level to log -// on your instance of logger, obtained with `logrus.New()`. -const ( - // PanicLevel level, highest level of severity. Logs and then calls panic with the - // message passed to Debug, Info, ... - PanicLevel Level = iota - // FatalLevel level. Logs and then calls `logger.Exit(1)`. It will exit even if the - // logging level is set to Panic. - FatalLevel - // ErrorLevel level. Logs. Used for errors that should definitely be noted. - // Commonly used for hooks to send errors to an error tracking service. - ErrorLevel - // WarnLevel level. Non-critical entries that deserve eyes. - WarnLevel - // InfoLevel level. General operational entries about what's going on inside the - // application. - InfoLevel - // DebugLevel level. Usually only enabled when debugging. Very verbose logging. - DebugLevel - // TraceLevel level. Designates finer-grained informational events than the Debug. - TraceLevel -) - -// Won't compile if StdLogger can't be realized by a log.Logger -var ( - _ StdLogger = &log.Logger{} - _ StdLogger = &Entry{} - _ StdLogger = &Logger{} -) - -// StdLogger is what your logrus-enabled library should take, that way -// it'll accept a stdlib logger and a logrus logger. There's no standard -// interface, this is the closest we get, unfortunately. -type StdLogger interface { - Print(...interface{}) - Printf(string, ...interface{}) - Println(...interface{}) - - Fatal(...interface{}) - Fatalf(string, ...interface{}) - Fatalln(...interface{}) - - Panic(...interface{}) - Panicf(string, ...interface{}) - Panicln(...interface{}) -} - -// The FieldLogger interface generalizes the Entry and Logger types -type FieldLogger interface { - WithField(key string, value interface{}) *Entry - WithFields(fields Fields) *Entry - WithError(err error) *Entry - - Debugf(format string, args ...interface{}) - Infof(format string, args ...interface{}) - Printf(format string, args ...interface{}) - Warnf(format string, args ...interface{}) - Warningf(format string, args ...interface{}) - Errorf(format string, args ...interface{}) - Fatalf(format string, args ...interface{}) - Panicf(format string, args ...interface{}) - - Debug(args ...interface{}) - Info(args ...interface{}) - Print(args ...interface{}) - Warn(args ...interface{}) - Warning(args ...interface{}) - Error(args ...interface{}) - Fatal(args ...interface{}) - Panic(args ...interface{}) - - Debugln(args ...interface{}) - Infoln(args ...interface{}) - Println(args ...interface{}) - Warnln(args ...interface{}) - Warningln(args ...interface{}) - Errorln(args ...interface{}) - Fatalln(args ...interface{}) - Panicln(args ...interface{}) - - // IsDebugEnabled() bool - // IsInfoEnabled() bool - // IsWarnEnabled() bool - // IsErrorEnabled() bool - // IsFatalEnabled() bool - // IsPanicEnabled() bool -} - -// Ext1FieldLogger (the first extension to FieldLogger) is superfluous, it is -// here for consistancy. Do not use. Use Logger or Entry instead. -type Ext1FieldLogger interface { - FieldLogger - Tracef(format string, args ...interface{}) - Trace(args ...interface{}) - Traceln(args ...interface{}) -} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_appengine.go b/vendor/github.com/sirupsen/logrus/terminal_check_appengine.go deleted file mode 100644 index 2403de9819..0000000000 --- a/vendor/github.com/sirupsen/logrus/terminal_check_appengine.go +++ /dev/null @@ -1,11 +0,0 @@ -// +build appengine - -package logrus - -import ( - "io" -) - -func checkIfTerminal(w io.Writer) bool { - return true -} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_bsd.go b/vendor/github.com/sirupsen/logrus/terminal_check_bsd.go deleted file mode 100644 index 499789984d..0000000000 --- a/vendor/github.com/sirupsen/logrus/terminal_check_bsd.go +++ /dev/null @@ -1,13 +0,0 @@ -// +build darwin dragonfly freebsd netbsd openbsd -// +build !js - -package logrus - -import "golang.org/x/sys/unix" - -const ioctlReadTermios = unix.TIOCGETA - -func isTerminal(fd int) bool { - _, err := unix.IoctlGetTermios(fd, ioctlReadTermios) - return err == nil -} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_js.go b/vendor/github.com/sirupsen/logrus/terminal_check_js.go deleted file mode 100644 index ebdae3ec62..0000000000 --- a/vendor/github.com/sirupsen/logrus/terminal_check_js.go +++ /dev/null @@ -1,7 +0,0 @@ -// +build js - -package logrus - -func isTerminal(fd int) bool { - return false -} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_no_terminal.go b/vendor/github.com/sirupsen/logrus/terminal_check_no_terminal.go deleted file mode 100644 index 97af92c68e..0000000000 --- a/vendor/github.com/sirupsen/logrus/terminal_check_no_terminal.go +++ /dev/null @@ -1,11 +0,0 @@ -// +build js nacl plan9 - -package logrus - -import ( - "io" -) - -func checkIfTerminal(w io.Writer) bool { - return false -} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_notappengine.go b/vendor/github.com/sirupsen/logrus/terminal_check_notappengine.go deleted file mode 100644 index 3293fb3caa..0000000000 --- a/vendor/github.com/sirupsen/logrus/terminal_check_notappengine.go +++ /dev/null @@ -1,17 +0,0 @@ -// +build !appengine,!js,!windows,!nacl,!plan9 - -package logrus - -import ( - "io" - "os" -) - -func checkIfTerminal(w io.Writer) bool { - switch v := w.(type) { - case *os.File: - return isTerminal(int(v.Fd())) - default: - return false - } -} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_solaris.go b/vendor/github.com/sirupsen/logrus/terminal_check_solaris.go deleted file mode 100644 index f6710b3bd0..0000000000 --- a/vendor/github.com/sirupsen/logrus/terminal_check_solaris.go +++ /dev/null @@ -1,11 +0,0 @@ -package logrus - -import ( - "golang.org/x/sys/unix" -) - -// IsTerminal returns true if the given file descriptor is a terminal. -func isTerminal(fd int) bool { - _, err := unix.IoctlGetTermio(fd, unix.TCGETA) - return err == nil -} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_unix.go b/vendor/github.com/sirupsen/logrus/terminal_check_unix.go deleted file mode 100644 index 04748b8515..0000000000 --- a/vendor/github.com/sirupsen/logrus/terminal_check_unix.go +++ /dev/null @@ -1,13 +0,0 @@ -// +build linux aix zos -// +build !js - -package logrus - -import "golang.org/x/sys/unix" - -const ioctlReadTermios = unix.TCGETS - -func isTerminal(fd int) bool { - _, err := unix.IoctlGetTermios(fd, ioctlReadTermios) - return err == nil -} diff --git a/vendor/github.com/sirupsen/logrus/terminal_check_windows.go b/vendor/github.com/sirupsen/logrus/terminal_check_windows.go deleted file mode 100644 index 2879eb50ea..0000000000 --- a/vendor/github.com/sirupsen/logrus/terminal_check_windows.go +++ /dev/null @@ -1,27 +0,0 @@ -// +build !appengine,!js,windows - -package logrus - -import ( - "io" - "os" - - "golang.org/x/sys/windows" -) - -func checkIfTerminal(w io.Writer) bool { - switch v := w.(type) { - case *os.File: - handle := windows.Handle(v.Fd()) - var mode uint32 - if err := windows.GetConsoleMode(handle, &mode); err != nil { - return false - } - mode |= windows.ENABLE_VIRTUAL_TERMINAL_PROCESSING - if err := windows.SetConsoleMode(handle, mode); err != nil { - return false - } - return true - } - return false -} diff --git a/vendor/github.com/sirupsen/logrus/text_formatter.go b/vendor/github.com/sirupsen/logrus/text_formatter.go deleted file mode 100644 index be2c6efe5e..0000000000 --- a/vendor/github.com/sirupsen/logrus/text_formatter.go +++ /dev/null @@ -1,339 +0,0 @@ -package logrus - -import ( - "bytes" - "fmt" - "os" - "runtime" - "sort" - "strconv" - "strings" - "sync" - "time" - "unicode/utf8" -) - -const ( - red = 31 - yellow = 33 - blue = 36 - gray = 37 -) - -var baseTimestamp time.Time - -func init() { - baseTimestamp = time.Now() -} - -// TextFormatter formats logs into text -type TextFormatter struct { - // Set to true to bypass checking for a TTY before outputting colors. - ForceColors bool - - // Force disabling colors. - DisableColors bool - - // Force quoting of all values - ForceQuote bool - - // DisableQuote disables quoting for all values. - // DisableQuote will have a lower priority than ForceQuote. - // If both of them are set to true, quote will be forced on all values. - DisableQuote bool - - // Override coloring based on CLICOLOR and CLICOLOR_FORCE. - https://bixense.com/clicolors/ - EnvironmentOverrideColors bool - - // Disable timestamp logging. useful when output is redirected to logging - // system that already adds timestamps. - DisableTimestamp bool - - // Enable logging the full timestamp when a TTY is attached instead of just - // the time passed since beginning of execution. - FullTimestamp bool - - // TimestampFormat to use for display when a full timestamp is printed. - // The format to use is the same than for time.Format or time.Parse from the standard - // library. - // The standard Library already provides a set of predefined format. - TimestampFormat string - - // The fields are sorted by default for a consistent output. For applications - // that log extremely frequently and don't use the JSON formatter this may not - // be desired. - DisableSorting bool - - // The keys sorting function, when uninitialized it uses sort.Strings. - SortingFunc func([]string) - - // Disables the truncation of the level text to 4 characters. - DisableLevelTruncation bool - - // PadLevelText Adds padding the level text so that all the levels output at the same length - // PadLevelText is a superset of the DisableLevelTruncation option - PadLevelText bool - - // QuoteEmptyFields will wrap empty fields in quotes if true - QuoteEmptyFields bool - - // Whether the logger's out is to a terminal - isTerminal bool - - // FieldMap allows users to customize the names of keys for default fields. - // As an example: - // formatter := &TextFormatter{ - // FieldMap: FieldMap{ - // FieldKeyTime: "@timestamp", - // FieldKeyLevel: "@level", - // FieldKeyMsg: "@message"}} - FieldMap FieldMap - - // CallerPrettyfier can be set by the user to modify the content - // of the function and file keys in the data when ReportCaller is - // activated. If any of the returned value is the empty string the - // corresponding key will be removed from fields. - CallerPrettyfier func(*runtime.Frame) (function string, file string) - - terminalInitOnce sync.Once - - // The max length of the level text, generated dynamically on init - levelTextMaxLength int -} - -func (f *TextFormatter) init(entry *Entry) { - if entry.Logger != nil { - f.isTerminal = checkIfTerminal(entry.Logger.Out) - } - // Get the max length of the level text - for _, level := range AllLevels { - levelTextLength := utf8.RuneCount([]byte(level.String())) - if levelTextLength > f.levelTextMaxLength { - f.levelTextMaxLength = levelTextLength - } - } -} - -func (f *TextFormatter) isColored() bool { - isColored := f.ForceColors || (f.isTerminal && (runtime.GOOS != "windows")) - - if f.EnvironmentOverrideColors { - switch force, ok := os.LookupEnv("CLICOLOR_FORCE"); { - case ok && force != "0": - isColored = true - case ok && force == "0", os.Getenv("CLICOLOR") == "0": - isColored = false - } - } - - return isColored && !f.DisableColors -} - -// Format renders a single log entry -func (f *TextFormatter) Format(entry *Entry) ([]byte, error) { - data := make(Fields) - for k, v := range entry.Data { - data[k] = v - } - prefixFieldClashes(data, f.FieldMap, entry.HasCaller()) - keys := make([]string, 0, len(data)) - for k := range data { - keys = append(keys, k) - } - - var funcVal, fileVal string - - fixedKeys := make([]string, 0, 4+len(data)) - if !f.DisableTimestamp { - fixedKeys = append(fixedKeys, f.FieldMap.resolve(FieldKeyTime)) - } - fixedKeys = append(fixedKeys, f.FieldMap.resolve(FieldKeyLevel)) - if entry.Message != "" { - fixedKeys = append(fixedKeys, f.FieldMap.resolve(FieldKeyMsg)) - } - if entry.err != "" { - fixedKeys = append(fixedKeys, f.FieldMap.resolve(FieldKeyLogrusError)) - } - if entry.HasCaller() { - if f.CallerPrettyfier != nil { - funcVal, fileVal = f.CallerPrettyfier(entry.Caller) - } else { - funcVal = entry.Caller.Function - fileVal = fmt.Sprintf("%s:%d", entry.Caller.File, entry.Caller.Line) - } - - if funcVal != "" { - fixedKeys = append(fixedKeys, f.FieldMap.resolve(FieldKeyFunc)) - } - if fileVal != "" { - fixedKeys = append(fixedKeys, f.FieldMap.resolve(FieldKeyFile)) - } - } - - if !f.DisableSorting { - if f.SortingFunc == nil { - sort.Strings(keys) - fixedKeys = append(fixedKeys, keys...) - } else { - if !f.isColored() { - fixedKeys = append(fixedKeys, keys...) - f.SortingFunc(fixedKeys) - } else { - f.SortingFunc(keys) - } - } - } else { - fixedKeys = append(fixedKeys, keys...) - } - - var b *bytes.Buffer - if entry.Buffer != nil { - b = entry.Buffer - } else { - b = &bytes.Buffer{} - } - - f.terminalInitOnce.Do(func() { f.init(entry) }) - - timestampFormat := f.TimestampFormat - if timestampFormat == "" { - timestampFormat = defaultTimestampFormat - } - if f.isColored() { - f.printColored(b, entry, keys, data, timestampFormat) - } else { - - for _, key := range fixedKeys { - var value interface{} - switch { - case key == f.FieldMap.resolve(FieldKeyTime): - value = entry.Time.Format(timestampFormat) - case key == f.FieldMap.resolve(FieldKeyLevel): - value = entry.Level.String() - case key == f.FieldMap.resolve(FieldKeyMsg): - value = entry.Message - case key == f.FieldMap.resolve(FieldKeyLogrusError): - value = entry.err - case key == f.FieldMap.resolve(FieldKeyFunc) && entry.HasCaller(): - value = funcVal - case key == f.FieldMap.resolve(FieldKeyFile) && entry.HasCaller(): - value = fileVal - default: - value = data[key] - } - f.appendKeyValue(b, key, value) - } - } - - b.WriteByte('\n') - return b.Bytes(), nil -} - -func (f *TextFormatter) printColored(b *bytes.Buffer, entry *Entry, keys []string, data Fields, timestampFormat string) { - var levelColor int - switch entry.Level { - case DebugLevel, TraceLevel: - levelColor = gray - case WarnLevel: - levelColor = yellow - case ErrorLevel, FatalLevel, PanicLevel: - levelColor = red - case InfoLevel: - levelColor = blue - default: - levelColor = blue - } - - levelText := strings.ToUpper(entry.Level.String()) - if !f.DisableLevelTruncation && !f.PadLevelText { - levelText = levelText[0:4] - } - if f.PadLevelText { - // Generates the format string used in the next line, for example "%-6s" or "%-7s". - // Based on the max level text length. - formatString := "%-" + strconv.Itoa(f.levelTextMaxLength) + "s" - // Formats the level text by appending spaces up to the max length, for example: - // - "INFO " - // - "WARNING" - levelText = fmt.Sprintf(formatString, levelText) - } - - // Remove a single newline if it already exists in the message to keep - // the behavior of logrus text_formatter the same as the stdlib log package - entry.Message = strings.TrimSuffix(entry.Message, "\n") - - caller := "" - if entry.HasCaller() { - funcVal := fmt.Sprintf("%s()", entry.Caller.Function) - fileVal := fmt.Sprintf("%s:%d", entry.Caller.File, entry.Caller.Line) - - if f.CallerPrettyfier != nil { - funcVal, fileVal = f.CallerPrettyfier(entry.Caller) - } - - if fileVal == "" { - caller = funcVal - } else if funcVal == "" { - caller = fileVal - } else { - caller = fileVal + " " + funcVal - } - } - - switch { - case f.DisableTimestamp: - fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m%s %-44s ", levelColor, levelText, caller, entry.Message) - case !f.FullTimestamp: - fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m[%04d]%s %-44s ", levelColor, levelText, int(entry.Time.Sub(baseTimestamp)/time.Second), caller, entry.Message) - default: - fmt.Fprintf(b, "\x1b[%dm%s\x1b[0m[%s]%s %-44s ", levelColor, levelText, entry.Time.Format(timestampFormat), caller, entry.Message) - } - for _, k := range keys { - v := data[k] - fmt.Fprintf(b, " \x1b[%dm%s\x1b[0m=", levelColor, k) - f.appendValue(b, v) - } -} - -func (f *TextFormatter) needsQuoting(text string) bool { - if f.ForceQuote { - return true - } - if f.QuoteEmptyFields && len(text) == 0 { - return true - } - if f.DisableQuote { - return false - } - for _, ch := range text { - if !((ch >= 'a' && ch <= 'z') || - (ch >= 'A' && ch <= 'Z') || - (ch >= '0' && ch <= '9') || - ch == '-' || ch == '.' || ch == '_' || ch == '/' || ch == '@' || ch == '^' || ch == '+') { - return true - } - } - return false -} - -func (f *TextFormatter) appendKeyValue(b *bytes.Buffer, key string, value interface{}) { - if b.Len() > 0 { - b.WriteByte(' ') - } - b.WriteString(key) - b.WriteByte('=') - f.appendValue(b, value) -} - -func (f *TextFormatter) appendValue(b *bytes.Buffer, value interface{}) { - stringVal, ok := value.(string) - if !ok { - stringVal = fmt.Sprint(value) - } - - if !f.needsQuoting(stringVal) { - b.WriteString(stringVal) - } else { - b.WriteString(fmt.Sprintf("%q", stringVal)) - } -} diff --git a/vendor/github.com/sirupsen/logrus/writer.go b/vendor/github.com/sirupsen/logrus/writer.go deleted file mode 100644 index 074fd4b8bd..0000000000 --- a/vendor/github.com/sirupsen/logrus/writer.go +++ /dev/null @@ -1,102 +0,0 @@ -package logrus - -import ( - "bufio" - "io" - "runtime" - "strings" -) - -// Writer at INFO level. See WriterLevel for details. -func (logger *Logger) Writer() *io.PipeWriter { - return logger.WriterLevel(InfoLevel) -} - -// WriterLevel returns an io.Writer that can be used to write arbitrary text to -// the logger at the given log level. Each line written to the writer will be -// printed in the usual way using formatters and hooks. The writer is part of an -// io.Pipe and it is the callers responsibility to close the writer when done. -// This can be used to override the standard library logger easily. -func (logger *Logger) WriterLevel(level Level) *io.PipeWriter { - return NewEntry(logger).WriterLevel(level) -} - -// Writer returns an io.Writer that writes to the logger at the info log level -func (entry *Entry) Writer() *io.PipeWriter { - return entry.WriterLevel(InfoLevel) -} - -// WriterLevel returns an io.Writer that writes to the logger at the given log level -func (entry *Entry) WriterLevel(level Level) *io.PipeWriter { - reader, writer := io.Pipe() - - var printFunc func(args ...interface{}) - - // Determine which log function to use based on the specified log level - switch level { - case TraceLevel: - printFunc = entry.Trace - case DebugLevel: - printFunc = entry.Debug - case InfoLevel: - printFunc = entry.Info - case WarnLevel: - printFunc = entry.Warn - case ErrorLevel: - printFunc = entry.Error - case FatalLevel: - printFunc = entry.Fatal - case PanicLevel: - printFunc = entry.Panic - default: - printFunc = entry.Print - } - - // Start a new goroutine to scan the input and write it to the logger using the specified print function. - // It splits the input into chunks of up to 64KB to avoid buffer overflows. - go entry.writerScanner(reader, printFunc) - - // Set a finalizer function to close the writer when it is garbage collected - runtime.SetFinalizer(writer, writerFinalizer) - - return writer -} - -// writerScanner scans the input from the reader and writes it to the logger -func (entry *Entry) writerScanner(reader *io.PipeReader, printFunc func(args ...interface{})) { - scanner := bufio.NewScanner(reader) - - // Set the buffer size to the maximum token size to avoid buffer overflows - scanner.Buffer(make([]byte, bufio.MaxScanTokenSize), bufio.MaxScanTokenSize) - - // Define a split function to split the input into chunks of up to 64KB - chunkSize := bufio.MaxScanTokenSize // 64KB - splitFunc := func(data []byte, atEOF bool) (int, []byte, error) { - if len(data) >= chunkSize { - return chunkSize, data[:chunkSize], nil - } - - return bufio.ScanLines(data, atEOF) - } - - // Use the custom split function to split the input - scanner.Split(splitFunc) - - // Scan the input and write it to the logger using the specified print function - for scanner.Scan() { - printFunc(strings.TrimRight(scanner.Text(), "\r\n")) - } - - // If there was an error while scanning the input, log an error - if err := scanner.Err(); err != nil { - entry.Errorf("Error while reading from Writer: %s", err) - } - - // Close the reader when we are done - reader.Close() -} - -// WriterFinalizer is a finalizer function that closes then given writer when it is garbage collected -func writerFinalizer(writer *io.PipeWriter) { - writer.Close() -} diff --git a/vendor/modules.txt b/vendor/modules.txt index b9becf20f9..017c1a3b7d 100644 --- a/vendor/modules.txt +++ b/vendor/modules.txt @@ -501,9 +501,6 @@ github.com/prometheus/procfs/internal/util # github.com/robfig/cron v1.2.0 ## explicit github.com/robfig/cron -# github.com/sirupsen/logrus v1.9.3 -## explicit; go 1.13 -github.com/sirupsen/logrus # github.com/spf13/cobra v1.9.1 ## explicit; go 1.15 github.com/spf13/cobra