Implements elastic search log downloader. Requires log-counter codex update!

This commit is contained in:
benbierens 2023-10-02 14:42:36 +02:00
parent c3ec64f25e
commit 6a96bd7639
No known key found for this signature in database
GPG Key ID: FE44815D96D0A1AA
6 changed files with 242 additions and 241 deletions

View File

@ -58,17 +58,6 @@ namespace Core
var workflow = entryPoint.Tools.CreateWorkflow();
return workflow.ExecuteCommand(container, command, args);
}
public IHttp CreateHttp()
{
var address = new Address("http://localhost", 9200);
var baseUrl = "";
return entryPoint.Tools.CreateHttp(address, baseUrl, client =>
{
// -H "kbn-xsrf: reporting" -H "Content-Type: application/json"
client.DefaultRequestHeaders.Add("kbn-xsrf", "reporting");
});
}
}
public interface IHasContainer

View File

@ -14,6 +14,7 @@ namespace Core
TResponse HttpPostJson<TRequest, TResponse>(string route, TRequest body);
string HttpPostJson<TRequest>(string route, TRequest body);
string HttpPostString(string route, string body);
TResponse HttpPostString<TResponse>(string route, string body);
string HttpPostStream(string route, Stream stream);
Stream HttpGetStream(string route);
T TryJsonDeserialize<T>(string json);
@ -99,6 +100,15 @@ namespace Core
}, $"HTTP-POST-STRING: {route}");
}
public TResponse HttpPostString<TResponse>(string route, string body)
{
var response = HttpPostString(route, body);
if (response == null) throw new Exception("Received no response.");
var result = JsonConvert.DeserializeObject<TResponse>(response);
if (result == null) throw new Exception("Failed to deserialize response");
return result;
}
public string HttpPostStream(string route, Stream stream)
{
return Retry(() =>

View File

@ -1,85 +0,0 @@
using Logging;
namespace ContinuousTests
{
public class ContainerLogStream
{
private readonly StreamReader reader;
private readonly Stream stream;
private readonly LogFile targetFile;
private readonly CancellationToken token;
private readonly TaskFactory taskFactory;
private int lastNumber = -1;
public bool Fault { get; private set; }
private bool run;
public ContainerLogStream(Stream stream, string name, LogFile targetFile, CancellationToken token, TaskFactory taskFactory)
{
this.stream = stream;
this.targetFile = targetFile;
this.token = token;
this.taskFactory = taskFactory;
Fault = false;
reader = new StreamReader(stream);
targetFile.Write(name);
}
public void Run()
{
run = true;
taskFactory.Run(() =>
{
while (run && !token.IsCancellationRequested)
{
Monitor();
}
});
}
public void Stop()
{
run = false;
stream.Close();
}
public void DeleteFile()
{
if (run) throw new Exception("Cannot delete file while stream is still running.");
File.Delete(targetFile.FullFilename);
}
private void Monitor()
{
var line = reader.ReadLine();
while (run && !string.IsNullOrEmpty(line) && !token.IsCancellationRequested)
{
ProcessLine(line);
line = reader.ReadLine();
}
}
private void ProcessLine(string s)
{
targetFile.WriteRaw(s);
// 000000004298
var sub = s.Substring(0, 12);
if (!int.TryParse(sub, out int number)) return;
if (lastNumber == -1)
{
lastNumber = number;
}
else
{
var expectedNumber = lastNumber + 1;
if (number != expectedNumber)
{
Fault = true;
}
lastNumber = number;
}
}
}
}

View File

@ -0,0 +1,210 @@
using Core;
using KubernetesWorkflow;
using Logging;
using Utils;
namespace ContinuousTests
{
public class ElasticSearchLogDownloader
{
private readonly IPluginTools tools;
private readonly ILog log;
public ElasticSearchLogDownloader(IPluginTools tools, ILog log)
{
this.tools = tools;
this.log = log;
}
public void Download(LogFile targetFile, RunningContainer container, DateTime startUtc, DateTime endUtc)
{
try
{
DownloadLog(targetFile, container, startUtc, endUtc);
}
catch (Exception ex)
{
log.Error("Failed to download log: " + ex);
}
}
private void DownloadLog(LogFile targetFile, RunningContainer container, DateTime startUtc, DateTime endUtc)
{
log.Log($"Downloading log (from ElasticSearch) for container '{container.Name}' within time range: " +
$"{startUtc.ToString("o")} - {endUtc.ToString("o")}");
var http = CreateElasticSearchHttp();
var queryTemplate = CreateQueryTemplate(container, startUtc, endUtc);
targetFile.Write($"Downloading '{container.Name}' to '{targetFile.FullFilename}'.");
var reconstructor = new LogReconstructor(targetFile, http, queryTemplate);
reconstructor.DownloadFullLog();
log.Log("Log download finished.");
}
private string CreateQueryTemplate(RunningContainer container, DateTime startUtc, DateTime endUtc)
{
var podName = container.Pod.PodInfo.Name;
var start = startUtc.ToString("o");
var end = endUtc.ToString("o");
var source = "{ \"sort\": [ { \"@timestamp\": { \"order\": \"asc\" } } ], \"fields\": [ { \"field\": \"@timestamp\", \"format\": \"strict_date_optional_time\" }, { \"field\": \"pod_name\" }, { \"field\": \"message\" } ], \"size\": <SIZE>, <SEARCHAFTER> \"_source\": false, \"query\": { \"bool\": { \"must\": [], \"filter\": [ { \"range\": { \"@timestamp\": { \"format\": \"strict_date_optional_time\", \"gte\": \"<STARTTIME>\", \"lte\": \"<ENDTIME>\" } } }, { \"match_phrase\": { \"pod_name\": \"<PODNAME>\" } } ] } } }";
return source
.Replace("<STARTTIME>", start)
.Replace("<ENDTIME>", end)
.Replace("<PODNAME>", podName);
}
private IHttp CreateElasticSearchHttp()
{
var address = new Address("http://localhost", 9200);
var baseUrl = "";
return tools.CreateHttp(address, baseUrl, client =>
{
client.DefaultRequestHeaders.Add("kbn-xsrf", "reporting");
});
}
public class LogReconstructor
{
private readonly List<LogQueueEntry> queue = new List<LogQueueEntry>();
private readonly LogFile targetFile;
private readonly IHttp http;
private readonly string queryTemplate;
private const int sizeOfPage = 2000;
private string searchAfter = "";
private int lastHits = 1;
private int lastLogLine = -1;
public LogReconstructor(LogFile targetFile, IHttp http, string queryTemplate)
{
this.targetFile = targetFile;
this.http = http;
this.queryTemplate = queryTemplate;
}
public void DownloadFullLog()
{
while (lastHits > 0)
{
QueryElasticSearch();
ProcessQueue();
}
}
private void QueryElasticSearch()
{
var query = queryTemplate
.Replace("<SIZE>", sizeOfPage.ToString())
.Replace("<SEARCHAFTER>", searchAfter);
var response = http.HttpPostString<SearchResponse>("_search", query);
lastHits = response.hits.hits.Length;
if (lastHits > 0)
{
UpdateSearchAfter(response);
foreach (var hit in response.hits.hits)
{
AddHitToQueue(hit);
}
}
}
private void AddHitToQueue(SearchHitEntry hit)
{
var message = hit.fields.message.Single();
var sub = message.Substring(0, 12);
if (int.TryParse(sub, out int number))
{
queue.Add(new LogQueueEntry(message, number));
}
}
private void UpdateSearchAfter(SearchResponse response)
{
var uniqueSearchNumbers = response.hits.hits.Select(h => h.sort.Single()).Distinct().ToList();
uniqueSearchNumbers.Reverse();
var searchNumber = GetSearchNumber(uniqueSearchNumbers);
searchAfter = $"\"search_after\": [{searchNumber}],";
}
private long GetSearchNumber(List<long> uniqueSearchNumbers)
{
if (uniqueSearchNumbers.Count == 1) return uniqueSearchNumbers.First();
return uniqueSearchNumbers.Skip(1).First();
}
private void ProcessQueue()
{
while (queue.Any())
{
var wantedNumber = lastLogLine + 1;
DeleteOldEntries(wantedNumber);
var currentEntry = queue.FirstOrDefault(e => e.Number == wantedNumber);
if (currentEntry != null)
{
WriteEntryToFile(currentEntry);
queue.Remove(currentEntry);
lastLogLine = currentEntry.Number;
}
else
{
// The line number we want is not in the queue.
// It will be returned by the elastic search query, some time in the future.
// Stop processing the queue for now.
return;
}
}
}
private void WriteEntryToFile(LogQueueEntry currentEntry)
{
targetFile.WriteRaw(currentEntry.Message);
}
private void DeleteOldEntries(int wantedNumber)
{
queue.RemoveAll(e => e.Number < wantedNumber);
}
public class LogQueueEntry
{
public LogQueueEntry(string message, int number)
{
Message = message;
Number = number;
}
public string Message { get; }
public int Number { get; }
}
public class SearchResponse
{
public SearchHits hits { get; set; }
}
public class SearchHits
{
public SearchHitEntry[] hits { get; set; }
}
public class SearchHitEntry
{
public SearchHitFields fields { get; set; }
public long[] sort { get; set; }
}
public class SearchHitFields
{
public string[] @timestamp { get; set; }
public string[] message { get; set; }
}
}
}
}

View File

@ -6,6 +6,7 @@ using System.Reflection;
using CodexPlugin;
using DistTestCore.Logs;
using Core;
using System.ComponentModel;
namespace ContinuousTests
{
@ -63,24 +64,18 @@ namespace ContinuousTests
private void RunTest(Action<bool> resultHandler)
{
var ci = entryPoint.CreateInterface();
var monitors = nodes.Select(n => new ContainerLogStream(
stream: ci.MonitorLog(n),
name: n.GetName(),
targetFile: fixtureLog.CreateSubfile(),
token: cancelToken,
taskFactory: taskFactory)).ToArray();
var testStart = DateTime.UtcNow;
try
{
foreach (var m in monitors) m.Run();
fixtureLog.Log("Monitor start");
RunTestMoments();
var duration = DateTime.UtcNow - testStart;
OverviewLog($" > Test passed. ({Time.FormatDuration(duration)})");
if (!config.KeepPassedTestLogs)
{
fixtureLog.Delete();
foreach (var m in monitors) m.DeleteFile();
}
resultHandler(true);
}
@ -89,6 +84,8 @@ namespace ContinuousTests
fixtureLog.Error("Test run failed with exception: " + ex);
fixtureLog.MarkAsFailed();
DownloadContainerLogs(testStart);
failureCount++;
resultHandler(false);
if (config.StopOnFailure > 0)
@ -101,12 +98,21 @@ namespace ContinuousTests
}
}
}
finally
}
private void DownloadContainerLogs(DateTime testStart)
{
// The test failed just now. We can't expect the logs to be available in elastic-search immediately:
Thread.Sleep(TimeSpan.FromMinutes(1));
var effectiveStart = testStart.Subtract(TimeSpan.FromSeconds(10));
var effectiveEnd = DateTime.UtcNow.AddSeconds(30);
var elasticSearchLogDownloader = new ElasticSearchLogDownloader(entryPoint.Tools, fixtureLog);
foreach (var node in nodes)
{
Thread.Sleep(1000);
fixtureLog.Log("Monitor stop");
foreach (var m in monitors) m.Stop();
if (monitors.Any(m => m.Fault)) throw new Exception("One or more downloaded container log is missing lines!");
var container = node.Container;
elasticSearchLogDownloader.Download(fixtureLog.CreateSubfile(), container, effectiveStart, effectiveEnd);
}
}
@ -144,7 +150,6 @@ namespace ContinuousTests
{
ThrowFailTest();
}
OverviewLog(" > Test passed.");
return;
}
}

View File

@ -2,9 +2,6 @@
using DistTestCore;
using GethPlugin;
using MetricsPlugin;
using Microsoft.IdentityModel.Abstractions;
using Microsoft.VisualStudio.TestPlatform.CommunicationUtilities;
using Newtonsoft.Json;
using NUnit.Framework;
using Utils;
@ -13,131 +10,6 @@ namespace Tests.BasicTests
[TestFixture]
public class ExampleTests : CodexDistTest
{
[Test]
public void AAA()
{
var http = this.Ci.CreateHttp();
//var query = "{\r\n \"query\": {\r\n \"bool\": {\r\n \"filter\": [\r\n {\r\n \"term\": {\r\n \"container_image.keyword\": \"docker.io/codexstorage/nim-codex:sha-9d735f9-dist-tests\"\r\n }\r\n },\r\n {\r\n \"term\": {\r\n \"pod_namespace.keyword\": \"codex-continuous-tests\"\r\n }\r\n },\r\n {\r\n \"term\": {\r\n \"pod_name.keyword\": \"codex3-workflow3-ff476767d-98zx4\"\r\n }\r\n },\r\n {\r\n \"range\": {\r\n \"@timestamp\": {\r\n \"lte\": \"2023-09-25T13:02:23.559Z\",\r\n \"gt\": \"2023-09-25T20:00:00.000Z\"\r\n }\r\n }\r\n }\r\n ]\r\n }\r\n }\r\n}";
//var query = "{ \"query\": {\"bool\": { \"filter\": [{ \"term\": {\"container_image.keyword\": \"docker.io/codexstorage/nim-codex:sha-9d735f9-dist-tests\" }},{ \"term\": {\"pod_namespace.keyword\": \"codex-continuous-tests\" }},{ \"term\": {\"pod_name.keyword\": \"codex3-workflow3-ff476767d-98zx4\" }},{ \"range\": {\"@timestamp\": { \"lte\": \"2023-09-29T13:02:23.559Z\", \"gt\": \"2023-09-20T20:00:00.000Z\"} }} ]} }}";
//var queryTemplate = "{ \"query\": {\"bool\": { \"filter\": [{ \"term\": {\"pod_namespace.keyword\": \"codex-continuous-tests\" }},{ \"term\": {\"pod_name.keyword\": \"bootstrap-2-599dfb4d65-v4v2b\" }},{ \"range\": {\"@timestamp\": { \"lte\": \"2023-10-03T13:02:23.559Z\", \"gt\": \"2023-09-01T20:00:00.000Z\"} }} ]} },\t\"_source\": \"message\",\t\"from\": <FROM>,\t\"size\": <SIZE>}";
var queryTemplate = "{ \"sort\": [ { \"@timestamp\": { \"order\": \"asc\" } } ], \"fields\": [ { \"field\": \"@timestamp\", \"format\": \"strict_date_optional_time\" }, { \"field\": \"pod_name\" }, { \"field\": \"message\" } ], \"size\": <SIZE>, <SEARCHAFTER> \"_source\": false, \"query\": { \"bool\": { \"must\": [], \"filter\": [ { \"range\": { \"@timestamp\": { \"format\": \"strict_date_optional_time\", \"gte\": \"2023-10-01T00:00:00.000Z\", \"lte\": \"2023-10-03T09:00:00.000Z\" } } }, { \"match_phrase\": { \"pod_name\": \"bootstrap-2-599dfb4d65-v4v2b\" } } ] } } }";
var outputFile = "c:\\Users\\Ben\\Desktop\\Cluster\\reconstructed.log";
var sizeOfPage = 2000;
var searchAfter = "";
var lastHits = 1;
var totalHits = 0;
var lastLogLine = -1;
var queue = new List<QueryLogEntry>();
//var jumpback = 0;
while (lastHits > 0)
{
//if (queue.Any()) jumpback++;
//else jumpback = 0;
var query = queryTemplate
.Replace("<SIZE>", sizeOfPage.ToString())
.Replace("<SEARCHAFTER>", searchAfter);
var output = http.HttpPostString("_search", query);
var response = JsonConvert.DeserializeObject<SearchResponse>(output)!;
lastHits = response.hits.hits.Length;
totalHits += response.hits.hits.Length;
if (lastHits > 0)
{
var uniqueSearchNumbers = response.hits.hits.Select(h => h.sort.Single()).Distinct().ToList();
uniqueSearchNumbers.Reverse();
var searchNumber = uniqueSearchNumbers.Skip(1).First().ToString();
searchAfter = $"\"search_after\": [{searchNumber}],";
foreach (var hit in response.hits.hits)
{
var message = hit.fields.message.Single();
var sub = message.Substring(0, 12);
if (int.TryParse(sub, out int number))
{
queue.Add(new QueryLogEntry(message, number));
}
}
}
// unload queue
var runQueue = 1;
while (runQueue > 0)
{
var wantedNumber = lastLogLine + 1;
var oldOnes = queue.Where(e => e.Number < wantedNumber).ToList();
foreach (var old in oldOnes) queue.Remove(old);
var entry = queue.FirstOrDefault(e => e.Number == wantedNumber);
if (entry != null)
{
File.AppendAllLines(outputFile, new[] { entry.Message });
queue.Remove(entry);
lastLogLine = entry.Number;
if (!queue.Any()) runQueue = 0;
}
else
{
runQueue = 0;
}
}
}
//var ouput = http.HttpGetString("");
var aaa = 0;
}
public class QueryLogEntry
{
public QueryLogEntry(string message, int number)
{
Message = message;
Number = number;
}
public string Message { get; }
public int Number { get; }
public override string ToString()
{
return Number.ToString();
}
}
public class SearchResponse
{
public SearchHits hits { get; set; }
}
public class SearchHits
{
public SearchHitEntry[] hits { get; set; }
}
public class SearchHitEntry
{
public SearchHitFields fields { get; set; }
public long[] sort { get; set; }
}
public class SearchHitFields
{
public string[] @timestamp { get; set; }
public string[] message { get; set; }
}
[Test]
public void CodexLogExample()
{