Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fluent bit gets stuck after "Bad file descriptor" error #3540

Closed
MPeli opened this issue May 24, 2021 · 6 comments
Closed

Fluent bit gets stuck after "Bad file descriptor" error #3540

MPeli opened this issue May 24, 2021 · 6 comments
Labels
Stale Windows Bugs and requests about Windows platforms

Comments

@MPeli
Copy link

MPeli commented May 24, 2021

Bug Report

Describe the bug
I often see an error saying "Bad file descriptor". Once this error shows up, fluent bit stops logging completely. Memory consumption increases significantly (up to 1GB).

To Reproduce
Not really sure how. I will add more details once I found more.

[error] [C:\projects\fluent-bit-2e87g\src\flb_http_client.c:1163 errno=9] Bad file descriptor
[2021/05/24 18:28:24] [ warn] [output:es:es.0] http_do=-1 URI=/_bulk

Expected behavior
Fluent bit should not get stuck and try to reconnect.

Your Environment

Fluent Bit Enterprise - SOS Report
==================================
The following report aims to be used by Fluent Bit and Fluentd community users.


[Fluent Bit]
    Version             1.7.6
    Built Flags          FLB_HAVE_PARSER FLB_HAVE_RECORD_ACCESSOR FLB_HAVE_STREAM_PROCESSOR JSMN_PARENT_LINKS JSMN_STRICT FLB_HAVE_TLS FLB_HAVE_AWS FLB_HAVE_SIGNV4 FLB_HAVE_SQLDB FLB_HAVE_TRACE FLB_HAVE_TIMESPEC_GET FLB_HAVE_PROXY_GO FLB_HAVE_REGEX FLB_HAVE_UTF8_ENCODER FLB_HAVE_LUAJIT

[Operating System]
    Name                Windows
    Version             6.2
    Build               9200

[Hardware]
    Architecture        x64 (AMD or Intel)
    Processors          12

[Built Plugins]
    Inputs              emitter tail dummy statsd storage_backlog stream_processor winlog tcp lib forward random
    Filters             alter_size aws record_modifier throttle throttle_size kubernetes modify nest parser expect grep rewrite_tag lua stdout geoip2
    Outputs             azure azure_blob counter datadog es file forward http influxdb logdna loki nrlogs null slack splunk stackdriver stdout syslog tcp flowcounter gelf websocket cloudwatch_logs kinesis_streams s3

[SERVER] Runtime configuration
    Flush               5.000000
    Daemon              On
    Log_Level           Trace

[INPUT] Instance
    Name                tail.0 (tail, id=0)
    Flags
    Threaded            No
    Tag                 <appname>-<pid>
    Mem_Buf_Limit       95.4M
    Path                C:\path\*_*.Flog3.log,D:\path\*_*.Flog3.x64.log
    Multiline           off
    Parser_Firstline    flog_parser
    Path_Key            path
    Offset_Key          1
    DB                  fluent.db
    Read_from_Head      On
    Tag_Regex           ^.*\\(?<appname>.+)_(?<pid>[0-9]+).Flog.*.log$
    Buffer_Chunk_Size   128k
    Buffer_Max_Size     256k
    Ignore_Older        10d
    Routes              es.0

[INPUT] Instance
    Name                storage_backlog.1 (storage_backlog, id=1)
    Flags
    Threaded            No
    Tag                 storage_backlog.1
    Routes              es.0


[OUTPUT] Instance
    Name                es.0 (es, id=0)
    Match               *
    TLS Active          Yes
    TLS.Verify          Off
    TLS.Ca_File         (not set)
    TLS.Crt_File        (not set)
    TLS.Key_File        (not set)
    TLS.Key_Passwd      (not set)
    Retry Limit         no limit
    Host.TCP_Port       443
    Host.Name           abcd.eu-west-1.es.amazonaws.com
    Index               fluent-bit
    Logstash_Format     true
    HTTP_User           aaa
    HTTP_Passwd         bbb
    Trace_Output        On
    Trace_Error         On
    Buffer_Size         False

Log files from three different machines
bad-file-descriptor-fluent-bit-I.log
bad-file-descriptor-fluent-bit-II.log
bad-file-descriptor-fluent-bit-III.log

@agup006 agup006 added the Windows Bugs and requests about Windows platforms label May 25, 2021
@fujimotos
Copy link
Member

I believe this is the same issue with #1022. There is a recovering issue in
the core engine (not only on WIndows) after some connection error.

@github-actions
Copy link
Contributor

github-actions bot commented Jul 4, 2021

This issue is stale because it has been open 30 days with no activity. Remove stale label or comment or this will be closed in 5 days.

@github-actions github-actions bot added the Stale label Jul 4, 2021
@pierluigilenoci
Copy link
Contributor

I've no hope of seeing it fixed.

@edsiper
Copy link
Member

edsiper commented Jul 5, 2021 via email

@github-actions github-actions bot removed the Stale label Jul 6, 2021
@github-actions
Copy link
Contributor

github-actions bot commented Aug 6, 2021

This issue is stale because it has been open 30 days with no activity. Remove stale label or comment or this will be closed in 5 days.

@github-actions github-actions bot added the Stale label Aug 6, 2021
@github-actions
Copy link
Contributor

This issue was closed because it has been stalled for 5 days with no activity.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Stale Windows Bugs and requests about Windows platforms
Projects
None yet
Development

No branches or pull requests

5 participants