Or any other log files/output? I'm open to any solution, but what I would like...
...is something where I can just click on a word or select some text and say "filter that out"
Something that colors different log levels differently, preferably automatically.
Something that can parse the "columns" and give me a nice quick list of values, like different unit names to filter out/solely include.
Something that lets me choose a time and go there. Something that lets me select only a specific timeframe of logs.
I know this can probably be done by going in/out of journalctl, recalling the last command and adding specific filter options... but it just feels slow. It's so many keypresses when I could just right click on the word and -> "Filter out/Search for" or something.
tbh my go to command is just.. journalctl -fe -u service
ex :
journalctl -fe -u jellyfin
journalctl -fe -u nordvpnd
so I'd also like to know the answer to this question. my other go to is dumping journalctl to text files and parsing with grep and awk and creating my own reports with that parsed information.
grep -E is my favorite, I love regex capturing groups.
I tried to setup Loki but the documentation was atrocious. Everything was outdated, referring to tools that were marked deprecated but documentation for the replacements just didn't exist.
It might be a bit overkill but I use Grafana to do this (with Loki). It's a pretty involved setup as well, but you can filter and search by content, or date/time. It's doable on a desktop but mainly servers use it
I don't know of any graphical tools that let you do this, but generally, if you want to search for specific terms/times/commands or anything of that sort, piping journalctl into grep (and optionally grep into less) is pretty effective at finding stuff.
In the server world that would usually involve doing something like sending the journal data to Elasticsearch using an Elasticsearch integration. But that involves setting up an Elasticsearch server and Kibana and so on which is very unwieldy for a desktop computer. It does work pretty well though in terms of filtering. But it also stores the data internally in indexes to speed up search.
Of course journald has a seemingly simple C API but writing code is a lot of work. There are probably API bindings for various languages.
Sounds like you want a siem like Wazuh. Its agent can collect journald logs from any number of systems. It also has a gui you can interact with to parse logs.
Well, just a monitoring stack, like for example Grafana, would probably be more suitable for this specific task (if we're doing central hosting/collection).
Kind of my main recommendation is to use something with OpenTelemetry. It's pretty much the standard protocol for transferring logs, traces and metrics, so if you set everything up with that, then you can swap out the visualization software with less pain.
Here's a guide for Grafana + OpenTelemetry Collector: https://grafana.com/docs/loki/latest/send-data/otel/
That is good but only shows the last 10-15 lines of the log, unless there is an arg to expand that, or a command to follow the log. I am aware of neither.
I usually use your suggested command to check if a service is up, then if it isn't, use journalctl to find out why.