-
Notifications
You must be signed in to change notification settings - Fork 40.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
structured logging for CSV format #43283
Comments
That's quite an interesting idea, I can see that this is useful. We have to do some research on existing CSV formats, i guess. |
My examples now:
Of course, you can union logfiles/metrics from multi instances, and it's very convenient. |
Hello, @mhalbritter! |
Hey @ivamly, thanks for the offer. For this issue, we'd like to spend some time on design work, so it's not open for contributions yet. |
@linux-china are you aware of any standards or conventions for the column ordering, the contents and their format, and so on in the CSV data for logging? |
Now I use @JsonPropertyOrder({"id", "nick", "email", "tags"})
public class User {
private Integer id;
private String nick;
private String email;
private String tags; For column type or format, and I think CSV Schema Language 1.2 some complicated. DuckDB use struct style: SELECT *
FROM read_csv('flights.csv',
delim = '|',
header = true,
columns = {
'FlightDate': 'DATE',
'UniqueCarrier': 'VARCHAR',
'OriginCityName': 'VARCHAR',
'DestCityName': 'VARCHAR'
}); For column ordering and format, and I think the following is fine.
type name is not required if it's text, for most time, and type is not necessary at all. Another question is about CSV headers. For new created logfile or rotatated logfile, the headers should be added as first line. |
This will be quite tricky for us as currently |
@philwebb CSV headers is not a must, and most developers will use schema by themselves, supplied by schema registry or input by themselves. |
Now structured logging built-in support for Elastic Common Schema (ecs), Graylog Extended Log Format (gelf) and Logstash (logstash), any plan to support CSV format? Now DuckDB, DataFusion, ClickHouse-local all support to query CSV directly with SQL, and CSV friendly to AWK/DataFrame too.
The text was updated successfully, but these errors were encountered: