You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
having Negative Lookup cache enabled error messages were shown
The exact command to reproduce the issue:
unclear, related active options are
performance.nl-cache-positive-entry: off
performance.nl-cache: on
performance.nl-cache-timeout: 60
The full output of the command that failed:
trigger is unclear, error message is like
[2024-07-02 20:23:31.323754 +0000] E [nl-cache-helper.c:229:__nlc_inode_clear_entries] (-->/usr/lib64/glusterfs/10.5/xlator/performance/nl-cache.so(+0x7470) [0x7ff0881b7470] -->/usr/lib64/glusterfs/10.5/xlator/performance/nl-cache.so(+0x73b5) [0x7ff0881b73b5] -->/usr/lib64/glusterfs/10.5/xlator/performance/nl-cache.so(+0x6cca) [0x7ff0881b6cca] ) 0-: Assertion failed: To attach gdb and coredump, Run the script under "glusterfs/extras/debug/gfcore.py
Expected results:
not such messges
Mandatory info: - The output of the gluster volume info command:
Volume Name: gfs-MOUNT
Type: Replicate
Volume ID: bb***
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: SERVER1:/gfs/brick-MOUNT
Brick2: SERVER2:/gfs/brick-MOUNT
Options Reconfigured:
features.scrub-freq: weekly
features.scrub-throttle: lazy
features.scrub: Active
features.bitrot: on
cluster.nufa: on
network.compression.compression-level: 5
network.compression.min-size: 1024
network.compression: off
auth.allow: 127.0.0.1,SERVER1-IP,SERVER2-IP
performance.client-io-threads: off
nfs.disable: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
cluster.granular-entry-heal: on
client.ssl: on
server.ssl: on
ssl.cipher-list: ECDHE-ECDSA-AES128-SHA256
performance.readdir-ahead: on
features.cache-invalidation: on
features.cache-invalidation-timeout: 600
performance.stat-prefetch: on
performance.cache-invalidation: on
cluster.rsync-hash-regex: on
performance.md-cache-timeout: 60
network.inode-lru-limit: 200000
performance.qr-cache-timeout: 60
performance.cache-size: 256MB
server.keepalive-count: 5
server.keepalive-interval: 2
server.keepalive-time: 10
performance.nl-cache-positive-entry: off
performance.nl-cache: on
performance.nl-cache-timeout: 60
- The output of the gluster volume status command:
(from SERVER2)
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick SERVER1:/gfs/brick-MOUNT 55858 0 Y 940
Brick SERVER2:/gfs/brick-MOUNT 59797 0 Y 889
Self-heal Daemon on localhost N/A N/A Y 1616
Bitrot Daemon on localhost N/A N/A Y 3914
Scrubber Daemon on localhost N/A N/A Y 4028
Self-heal Daemon on SERVER1 N/A N/A Y 2711
Bitrot Daemon on SERVER1 N/A N/A Y 2572
Scrubber Daemon on SERVER1 N/A N/A Y 2629
Task Status of Volume gfs-MOUNT
------------------------------------------------------------------------------
There are no active volume tasks
- The output of the gluster volume heal command:
Launching heal operation to perform index self heal on volume gfs-MOUNT has been successful
Use heal info commands to check status.
- Provide logs present on following locations of client and server nodes -
On request, error log message already shown above
- Is there any crash ? Provide the backtrace and coredump
Not seen.
Additional info:
Message disappeared once performance.nl-cache was set to off
performance.nl-cache-positive-entryon or off has no influcence
- The operating system / glusterfs version:
system: EL 9.4
glusterfs: 10.5
The text was updated successfully, but these errors were encountered:
Description of problem:
having Negative Lookup cache enabled error messages were shown
The exact command to reproduce the issue:
unclear, related active options are
The full output of the command that failed:
trigger is unclear, error message is like
Expected results:
not such messges
Mandatory info:
- The output of the
gluster volume info
command:- The output of the
gluster volume status
command:(from SERVER2)
- The output of the
gluster volume heal
command:- Provide logs present on following locations of client and server nodes -
On request, error log message already shown above
- Is there any crash ? Provide the backtrace and coredump
Not seen.
Additional info:
Message disappeared once
performance.nl-cache
was set tooff
performance.nl-cache-positive-entry
on
oroff
has no influcence- The operating system / glusterfs version:
system: EL 9.4
glusterfs: 10.5
The text was updated successfully, but these errors were encountered: