is anyone out there using s3 input? and if so, how are you dealing with the fact that @timestamp is off by ~15 mins (due to delayed log delivery to bucket by aws)? i thought of updating @timestamp with value of actual event. but then i read that jordan sissel recommends not updating @timestmp as it could have unexpected results
nmische_ has quit
nmische_ joined the channel
whack
maseda: I don't thinK I've ever recommended that
you should always try to use the timestamp in the logs themselves if you can
javascott2 joined the channel
javascott has quit
eper
this jordan sissel guy sounds like he's never used logstash ho ho ho
wonder if he's any relation to the author v.similar names
fedesilv_ has quit
fedesilva joined the channel
maseda
eper: heh. i misquoted him. he said dont use mutate to update @timestamp
eper
oh yeah always use date {}
bline79
has anyone experienced this "Starting Elasticsearch Server:sysctl: permission denied on key 'vm.max_map_count'" with openvz ?
whack
maseda: yeah, if you try to use mutate to set @timestamp now, you will get an error.
maseda
whack: i had trouble using log-provided timestamp since field was string versus date type in ES
whack: roger that.
spinscale has quit
clay584 has quit
whack: i previously tried range searches on my the timestamps of actual event and ES complained
whack: maybe my query was hosed though. is it possible to range search on ES string field?
whack
sounds unrelated, what was the error?
yes, you can raange search strings
maseda
whack: crapola. i must have screwed up my query then. i will try again. blessings to you
nicholasf has quit
freezey joined the channel
whack: does the date need to be in a special format though? i was using dates from tomcat like "2014-01-21 13:41:28,024"
stackedsax has quit
whack
maseda: i'm confused. Are we talking strings or dates, and with elasticsearch, right?
stackedsax joined the channel
knoxilla has quit
cwebber joined the channel
maseda
whack: yes this is purely ES question at this point. i wanted to search like this: "url_access_time:[now-1h TO now]". that blew up on me. but then i tried similar search on @timestamp and it was successful. then i looked at ES mappings and low and behold url_access_time was string and @timestamp was date.
whack
right
you can't do time searches on strings
you can only do time searches on dates ;)
nicholasf joined the channel
nemish
whack: quick question... if necessary would it be an issue load balancing the TCP inputs of logstash behind a hardware load balancer? I can't think of any since each connection is an individual transaction
whack
nemish: "each connection is an individual transaction" ? You only send one log per connection?
frankS2
eper: do you have any idea how to add a field for sender id for your exim filter?
weird nobody have written a complete exim filter yet :P
warkolm joined the channel
nemish
whack: true... it could be batch of events... but you get the point... there shouldn't be any issue sending to 2 LS servers sitting behind a LB right? the LB locks each connection... i could make them sticky if need be I guess
matanya_ is now known as matanya
whack
nope, it should be fine
you don't need to make them sticky
maseda
whack: i might have misundertood you before... when you said "always try to use the timestamp in the logs themselves if you can", you meant @timestamp right?
whack
maseda: you are pulling logfiles from s3, yes?
eper
frankS2: not of the top of my head, tried hard to forget the format of exim logs as they make me a sad panda
maseda
whack: correct
whack
maseda: there's some weird timestamp format *in* that file, yes?
maseda
whack: tomcat as well
whack: correct
whack
maseda: you want to use the timestamp *in* the tomcat log as the @timestamp
nemish
whack: thanks... if i had LS running on all hosts i probably would just LB redis... but since I use many different methods of getting into LS as the aggregate... then REDIS... and then LS pops off REDIS to ES
whack
you achieve this with the date filter
maseda
whack: ahh... i am having problems with that.
Scissor has quit
Scissor joined the channel
whack: i cant seem to find correct pseudocode. do i just need match and that is it?
jedi4ever has quit
nemish
whack: so you made a featured appearance in the marvel video huh? ;)
maseda
whack: the other problem is that s3 log format has ugly apache-style date and i wanted to keep JODA
nicholasf has quit
whack
nemish: hehe
nemish
rashidkpc: any special requirements for load balancing a bunch of kibana web frontends? sticky necessary?
whack
maseda: you use grok to match the date text from your log, then use the date filter to parse it
nemish: no, kibana is stateless
in fact, kibana is literally just static html and javascript
vanberge1 has quit
nemish
whack: cool... so 6 kibana web heads it is :P
eper
can not see much use in load balancing kibana unless its for HA
whack
nemish: I doubt you need that.
nemish
whack: more the merrier ;)
whack
nemish: once kibana is served (say, from apache), it's unlikely to hit apache again.
pheaver has quit
mr-potter joined the channel
it hits elasticsearch for search, but nothing else after, iirc
pheaver joined the channel
vanberge1 joined the channel
maseda
whack: i have done the grok. i have extracted out %{MONTHDAY}/%{MONTH}/%{YEAR}:%{TIME} %{INT:TIMEOFFSET} with it. i guess you are saying that i just put syntax similar to this in match option for date filter and then @timestamp will show as "06/Jan/2014:13:51:25 +0000" for example? if so, will @timestamp still be range-searchable with that ugly apache-style date format?
whack
in logstash 1.2.2 and beyond, you can set a 'target' for the date format if you need multiple date fields
but yes
you can take your apache log timestamp format, parse it and store it in @timestamp using the date filter
which will let you do time range searches
nicholasf joined the channel
davuxx has quit
davuxx joined the channel
csd126
whack: Yesterday you made a change to allow multiple geoip filters. I tested that today and it works, but I'm running into the infamous "UndefinedConversionError: ""\xF3"" from ASCII-8BIT to UTF-8" errors now. That was the only change I introduced so I'm guessing it has something to do with that plugin. I need to do some experimentation to even have enough to write a trouble ticket.
Anton_ has quit
ehaselwanter has quit
Just wanted to give you a heads-up in case you could see something obvious on the geoip filter side of things.
paulczar is now known as zz_paulczar
whack
csd126: can you give me an IP lookup that results in that?
csd126: would love to see where the bad dat comes from; gotta force all the crap from geoip to be UTF-8
nmische__ joined the channel
nmische_ has quit
csd126
whack: Sure, give me a minute to see if I can find one.
michaelhart has quit
kalloc joined the channel
jkitchen
question!
eper
hmm
on TCP input %{host} becomes IP:PORT on UDP input %{host} is just IP
jkitchen
logstash applies things in the order it finds them in the config file. does this also apply to inputs and outputs? or are inputs effectively *always* before *all* filters and outputs are always effectively after all filters?
eper
that must be a bug O_o
Anton_ joined the channel
whack
eper: bug.
eper
*stops the tcp rsyslog rollout*
jkitchen
like if I have file A with input, filter output... and file B with input filter output, will the filter and output from file A also be applied to the input in file B?
eper
:p
whack
jkitchen: logically speaking, in a pipeline, you can't have an input happen *after* a filter
kalloc_ joined the channel
kalloc has quit
moos3
is there away to pump flat file logs from windows into logstash nicely ?
zz_paulczar is now known as paulczar
jlawson_ has quit
ehthayer has quit
csd126
whack: Do I need to have elasticsearch in my pipeline to see these UTF-8 errors?
(right now I'm just displaying them to stdout)
kalloc_ has quit
itsted joined the channel
rcleere has quit
chrisgilmerproj has left the channel
matanya has quit
ronnocol
moving multiline from LS to beaver has made life better... now I don't need a second event to kick the first one out of the queue.
techminer has quit
techminer joined the channel
libtek1 joined the channel
princessleah joined the channel
moos3
when running logstash in agent mode, is there a good limit for memory ? I'm thinking of using it on some web servers and want to pipe php, httpd access/error logs to logstash indexer
mveitas has quit
kjstone00 joined the channel
danshultz has quit
danshultz joined the channel
jotterbot1234 joined the channel
rcleere joined the channel
JasonG joined the channel
nmische__ has quit
itsted
grok help please: i have a pattern that "works" in grokdebug, but breaks because of a double quote. (a log line is from a tomcat access log and begins like this: 10.189.254.4 - - [30/Jan/2014:15:06:15 -0500] "POST /foo... I want to parse out the http verb separately, so I don't want to use a QS. Escaping the double quote doesn't seem to work. Seems like this should be really simple. Thoughts?
stackedsax has quit
whack
itsted: try single qutoes around the pattern instead of double