Skip to content

Commit

Permalink
allow for json/ndjson content type with charset (#32767)
Browse files Browse the repository at this point in the history
* allow for json/ndjson content type with charset

* add '; charset=UTF-8' in integration tests

* changelog

* assessing CI

* revert correct behaviour

(cherry picked from commit b40349c)
  • Loading branch information
Andrea Spacca authored and mergify[bot] committed Aug 25, 2022
1 parent 1710855 commit 2396f8a
Show file tree
Hide file tree
Showing 3 changed files with 18 additions and 8 deletions.
5 changes: 5 additions & 0 deletions CHANGELOG.next.asciidoc
Original file line number Diff line number Diff line change
Expand Up @@ -43,6 +43,11 @@ https://github.com/elastic/beats/compare/v8.2.0\...main[Check the HEAD diff]

*Filebeat*

- Fix counter for number of events published in `httpjson` input. {pull}31993[31993]
- Fix handling of Checkpoint event for R81. {issue}32380[32380] {pull}32458[32458]
- Fix a hang on `apt-get update` stage in packaging. {pull}32580[32580]
- gcp-pubsub input: Restart Pub/Sub client on all errors. {issue}32550[32550] {pull}32712[32712]
- Fix not parsing as json when `json` and `ndjson` content types have charset information in `aws-s3` input {pull}32767[32767]
- Update `cloud.region` parsing in cloudtrail fileset. {pull}32763[32763]
- Fix file.path field in cloudtrail fileset to use json.digestS3Object. {pull}32759[32759]

Expand Down
19 changes: 12 additions & 7 deletions x-pack/filebeat/input/awss3/input_integration_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -18,6 +18,7 @@ import (
"path"
"path/filepath"
"runtime"
"strings"
"testing"
"time"

Expand Down Expand Up @@ -88,7 +89,6 @@ file_selectors:
-
regex: 'events-array.json$'
expand_event_list_from_field: Events
content_type: application/json
include_s3_metadata:
- last-modified
- x-amz-version-id
Expand All @@ -97,7 +97,6 @@ file_selectors:
- Content-Type
-
regex: '\.(?:nd)?json(\.gz)?$'
content_type: application/json
-
regex: 'multiline.txt$'
parsers:
Expand All @@ -117,7 +116,6 @@ file_selectors:
-
regex: 'events-array.json$'
expand_event_list_from_field: Events
content_type: application/json
include_s3_metadata:
- last-modified
- x-amz-version-id
Expand All @@ -126,7 +124,6 @@ file_selectors:
- Content-Type
-
regex: '\.(?:nd)?json(\.gz)?$'
content_type: application/json
-
regex: 'multiline.txt$'
parsers:
Expand Down Expand Up @@ -328,11 +325,19 @@ func uploadS3TestFiles(t *testing.T, region, bucket string, filenames ...string)
t.Fatalf("Failed to open file %q, %v", filename, err)
}

contentType := ""
if strings.HasSuffix(filename, "ndjson") || strings.HasSuffix(filename, "ndjson.gz") {
contentType = contentTypeNDJSON + "; charset=UTF-8"
} else if strings.HasSuffix(filename, "json") || strings.HasSuffix(filename, "json.gz") {
contentType = contentTypeJSON + "; charset=UTF-8"
}

// Upload the file to S3.
result, err := uploader.Upload(context.Background(), &s3.PutObjectInput{
Bucket: aws.String(bucket),
Key: aws.String(filepath.Base(filename)),
Body: bytes.NewReader(data),
Bucket: aws.String(bucket),
Key: aws.String(filepath.Base(filename)),
Body: bytes.NewReader(data),
ContentType: aws.String(contentType),
})
if err != nil {
t.Fatalf("Failed to upload file %q: %v", filename, err)
Expand Down
2 changes: 1 addition & 1 deletion x-pack/filebeat/input/awss3/s3_objects.go
Original file line number Diff line number Diff line change
Expand Up @@ -153,7 +153,7 @@ func (p *s3ObjectProcessor) ProcessS3Object() error {

// Process object content stream.
switch {
case contentType == contentTypeJSON || contentType == contentTypeNDJSON:
case strings.HasPrefix(contentType, contentTypeJSON) || strings.HasPrefix(contentType, contentTypeNDJSON):
err = p.readJSON(reader)
default:
err = p.readFile(reader)
Expand Down

0 comments on commit 2396f8a

Please sign in to comment.