You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Since I don't use that feature never noticed this huge issue. After reading tcpdump manual it seems we can use the -C flag which says:
-C file_size
Before writing a raw packet to a savefile, check whether the file is currently larger than file_size and, if so,
close the current savefile and open a new one. Savefiles after the first savefile will have the name specified
with the -w flag, with a number after it, starting at 1 and continuing upward. The units of file_size are mil‐
lions of bytes (1,000,000 bytes, not 1,048,576 bytes).
Then, check for old pcap files each time the cloud flare ip scanner runs and delete them.
I ended up getting a 2tb pcap file and crashing my server due to this....
The text was updated successfully, but these errors were encountered: