mirror of
https://github.com/elastic/logstash.git
synced 2025-04-25 07:07:54 -04:00
- swear less, and never in documentation.
This commit is contained in:
parent
39b31baca6
commit
f1d6d1d71c
2 changed files with 1 additions and 6 deletions
|
@ -55,10 +55,6 @@ require "logstash/namespace"
|
||||||
# TODO If you have bugs report or helpful advice contact me, but remember that this code is much mine as much as yours,
|
# TODO If you have bugs report or helpful advice contact me, but remember that this code is much mine as much as yours,
|
||||||
# try to work on it if you want :)
|
# try to work on it if you want :)
|
||||||
|
|
||||||
# The programmer's question is: "Why you fuck you use name ls.s3.... you kidding me, motherfucker?
|
|
||||||
# The answer is simple, s3 not allow special characters like "/" "[,]", very useful in date format,
|
|
||||||
# because if you use them s3 dosen't know no more the key and send you to hell!
|
|
||||||
# For example "/" in s3 means you can specify a subfolder on bucket.
|
|
||||||
|
|
||||||
# USAGE:
|
# USAGE:
|
||||||
|
|
||||||
|
@ -220,7 +216,6 @@ class LogStash::Outputs::S3 < LogStash::Outputs::Base
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@logger.debug "S3: let's destroying the temporary shit file "+name_file
|
|
||||||
File.delete (file)
|
File.delete (file)
|
||||||
|
|
||||||
end
|
end
|
||||||
|
|
|
@ -12,7 +12,7 @@ describe "parse mysql slow query log", :if => false do
|
||||||
# I have seen) a 'SET timestamp=123456789' line in each query event, so
|
# I have seen) a 'SET timestamp=123456789' line in each query event, so
|
||||||
# I use that as the timestamp instead.
|
# I use that as the timestamp instead.
|
||||||
#
|
#
|
||||||
# mysql logs are fucked up, so this is pretty much best effort.
|
# mysql logs are messed up, so this is pretty much best effort.
|
||||||
match => [ "@message", "^# Time: " ]
|
match => [ "@message", "^# Time: " ]
|
||||||
negate => true
|
negate => true
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue