Realized I never replied-all... Re-adding users_list On Mon, Aug 15, 2016 at 10:58 AM, Eric Wolinetz <ewoli...@redhat.com> wrote:
> Fluentd tries to connect to both "logging-es" | "logging-es-ops" in the > logging namespace (if you're using the ops deployment) and "kubernetes" in > the default namespace. I think in this case it is having trouble > connecting to the kubernetes service to look up metadata for your > containers. > > > On Mon, Aug 15, 2016 at 10:54 AM, Frank Liauw <fr...@vsee.com> wrote: > >> Oh stupid me; I was confused by my own namespaces; was looking at the >> wrong namespace, thinking that's the one with pods that have an active log >> stream. The logs are ingested fine, thanks for your assistance! :) >> >> On the possible DNS issue of fluentd on one of my nodes, what hostname is >> fluentd trying to reach when starting up? We did perform some network >> changes to this particular node to aid public routing, but as far as the >> routing table is concerned, it should not have made a difference for local >> traffic. >> >> Normal functioning node without public routing changes >> >> [root@node1 network-scripts]# route -n >> Kernel IP routing table >> Destination Gateway Genmask Flags Metric Ref Use >> Iface >> 0.0.0.0 10.10.0.5 0.0.0.0 UG 100 0 0 >> ens160 >> 10.1.0.0 0.0.0.0 255.255.0.0 U 0 0 0 >> tun0 >> 10.10.0.0 0.0.0.0 255.255.0.0 U 100 0 0 >> ens160 >> 172.30.0.0 0.0.0.0 255.255.0.0 U 0 0 0 >> tun0 >> >> Malfunctioning node with public routing changes >> >> [root@node2 network-scripts]# route -n >> Kernel IP routing table >> Destination Gateway Genmask Flags Metric Ref Use >> Iface >> 0.0.0.0 199.27.105.1 0.0.0.0 UG 100 0 0 >> ens192 >> 10.0.0.0 10.10.0.5 255.0.0.0 UG 100 0 0 >> ens160 >> 10.1.0.0 0.0.0.0 255.255.0.0 U 0 0 0 >> tun0 >> 10.10.0.0 0.0.0.0 255.255.0.0 U 100 0 0 >> ens160 >> 172.30.0.0 0.0.0.0 255.255.0.0 U 0 0 0 >> tun0 >> 199.27.105.0 0.0.0.0 255.255.255.128 U 100 0 0 >> ens192 >> >> Frank >> Systems Engineer >> >> VSee: fr...@vsee.com <http://vsee.com/u/tmd4RB> | Cell: +65 9338 0035 >> >> Join me on VSee for Free <http://vsee.com/u/tmd4RB> >> >> >> >> >> On Mon, Aug 15, 2016 at 11:23 PM, Eric Wolinetz <ewoli...@redhat.com> >> wrote: >> >>> Correct, the way Fluentd pulls in the logs for your other containers is >>> the same pipeline used for collecting logs for the below shown Kibana pod. >>> >>> Going back to your ES logs, can you verify the date portion of a >>> microsvc index line? >>> We can then update time range in the upper-right corner of Kibana to >>> change from the last hour to something like the last month (something that >>> would encompass the date for the index). >>> >>> >>> On Mon, Aug 15, 2016 at 10:15 AM, Frank Liauw <fr...@vsee.com> wrote: >>> >>>> Screencap is as follows: >>>> >>>> >>>> The query is as simple as it gets, *. I see my namespaces / projects as >>>> indexes. >>>> >>>> I see logs for logging project just fine: >>>> >>>> >>>> >>>> Fluentd is not ingesting the logs for pods in my namespaces. I'm yet to >>>> pull apart how fluentd does that, though there's no reason why logs for my >>>> other pods aren't getting indexed whereas kibana logs are if they are both >>>> ingested by fluentd, assuming that kibana logs use the same pipeline as all >>>> other pod logs. >>>> >>>> Frank >>>> Systems Engineer >>>> >>>> VSee: fr...@vsee.com <http://vsee.com/u/tmd4RB> | Cell: +65 9338 0035 >>>> >>>> Join me on VSee for Free <http://vsee.com/u/tmd4RB> >>>> >>>> >>>> >>>> >>>> On Mon, Aug 15, 2016 at 10:59 PM, Eric Wolinetz <ewoli...@redhat.com> >>>> wrote: >>>> >>>>> Can you either send a screencap of your Kibana console? Or describe >>>>> how you are accessing Kibana and what you are seeing? (e.g. your query >>>>> string, the index you're querying on, the time range for fetched >>>>> responses) >>>>> >>>>> On Mon, Aug 15, 2016 at 9:55 AM, Frank Liauw <fr...@vsee.com> wrote: >>>>> >>>>>> I can see indexes of my namespaces, but nothing going on in actual >>>>>> logs in kibana though. >>>>>> >>>>>> Frank >>>>>> Systems Engineer >>>>>> >>>>>> VSee: fr...@vsee.com <http://vsee.com/u/tmd4RB> | Cell: +65 9338 0035 >>>>>> >>>>>> Join me on VSee for Free <http://vsee.com/u/tmd4RB> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Mon, Aug 15, 2016 at 10:37 PM, Eric Wolinetz <ewoli...@redhat.com> >>>>>> wrote: >>>>>> >>>>>>> True, we should be able to. You should be able to see entries in >>>>>>> the master ES node's logs that indices were created. Based on your log >>>>>>> snippet it should be "One Above All" in this pod: >>>>>>> logging-es-0w45va6n-2-8m8 >>>>>>> 5p >>>>>>> >>>>>>> If we don't see anything like "[logging.iasdf9123j-asdf.2016.02.19] >>>>>>> creating index" then there might be something else going on... Do you >>>>>>> see >>>>>>> any entries like that for your projects? >>>>>>> >>>>>>> On Mon, Aug 15, 2016 at 9:05 AM, Frank Liauw <fr...@vsee.com> wrote: >>>>>>> >>>>>>>> I reckoned the same, and already tried doing that, but still got >>>>>>>> the same message. That node have the same configuration as all my other >>>>>>>> nodes, so I don't see why that's the case. >>>>>>>> >>>>>>>> Even if it is having DNS issues, I should still be seeing logs from >>>>>>>> all other nodes that didn't report that error? >>>>>>>> >>>>>>>> Frank >>>>>>>> Systems Engineer >>>>>>>> >>>>>>>> VSee: fr...@vsee.com <http://vsee.com/u/tmd4RB> | Cell: +65 9338 >>>>>>>> 0035 >>>>>>>> >>>>>>>> Join me on VSee for Free <http://vsee.com/u/tmd4RB> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Mon, Aug 15, 2016 at 9:49 PM, Eric Wolinetz <ewoli...@redhat.com >>>>>>>> > wrote: >>>>>>>> >>>>>>>>> It looks like that one Fluentd pod is having DNS issues making a >>>>>>>>> connection to the Kubernetes service. Can you try deleting that pod >>>>>>>>> to >>>>>>>>> have a new one redeployed and see if that is in a better state? >>>>>>>>> >>>>>>>>> On Mon, Aug 15, 2016 at 2:54 AM, Frank Liauw <fr...@vsee.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi All, >>>>>>>>>> >>>>>>>>>> I followed through the instructions on >>>>>>>>>> https://docs.openshift.org/latest/install_config/aggregat >>>>>>>>>> e_logging.html and have setup a 3 node ES cluster. Fluentd is >>>>>>>>>> also deployed on all my nodes. >>>>>>>>>> >>>>>>>>>> I am getting kibana logs on the logging project, but all my other >>>>>>>>>> projects do not have any logs; kibana shows "No results found", with >>>>>>>>>> occasional errors reading "Discover: An error occurred with your >>>>>>>>>> request. >>>>>>>>>> Reset your inputs and try again." >>>>>>>>>> >>>>>>>>>> Probing the requests made by kibana, some calls to >>>>>>>>>> /elasticsearch/_msearch?timeout=0&ignore_unavailable=true >>>>>>>>>> &preference=1471245075265 are failing from time to time. >>>>>>>>>> >>>>>>>>>> Looking into the ES logs for all 3 cluster pods, I don't see much >>>>>>>>>> errors to be concerned, with the last error of 2 nodes similar to the >>>>>>>>>> following which seems to be a known issue with Openshift's setup ( >>>>>>>>>> https://lists.openshift.redhat.com/openshift-archives/users >>>>>>>>>> /2015-December/msg00078.html) and possibly explains the failed >>>>>>>>>> requests made by kibana on auto-refresh, but that's a problem for >>>>>>>>>> another >>>>>>>>>> day: >>>>>>>>>> >>>>>>>>>> [2016-08-15 06:53:49,130][INFO ][cluster.service ] >>>>>>>>>> [Gremlin] added {[Quicksilver][t2l6Oz8uT-WS8Fa >>>>>>>>>> 7S7jzfQ][logging-es-d7r1t3dm-2-a0cf0][inet[/10.1.3.3:9300]],}, >>>>>>>>>> reason: zen-disco-receive(from master [[One Above >>>>>>>>>> All][CyFgyTTtS_S85yYRom2wVQ][logging-es-0w45va6n-2-8m85p][in >>>>>>>>>> et[/10.1.2.5:9300]]]) >>>>>>>>>> [2016-08-15 06:59:27,727][ERROR][com.flora >>>>>>>>>> gunn.searchguard.filter.SearchGuardActionFilter] Error while >>>>>>>>>> apply() due to com.floragunn.searchguard.toke >>>>>>>>>> neval.MalformedConfigurationException: no bypass or execute >>>>>>>>>> filters at all for action indices:admin/mappings/fields/get >>>>>>>>>> com.floragunn.searchguard.tokeneval.MalformedConfigurationException: >>>>>>>>>> no bypass or execute filters at all >>>>>>>>>> >>>>>>>>>> Looking into fluentd logs, one of my nodes is complaining of a >>>>>>>>>> "getaddrinfo" error: >>>>>>>>>> >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: unexpected error >>>>>>>>>> error="getaddrinfo: Name or service not known" >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/ruby/net/http.rb:878:in `initialize' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/ruby/net/http.rb:878:in `open' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/ruby/net/http.rb:878:in `block in connect' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/ruby/timeout.rb:52:in `timeout' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/ruby/net/http.rb:877:in `connect' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/ruby/net/http.rb:862:in `do_start' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/ruby/net/http.rb:851:in `start' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/rest-client-2.0.0/lib/restclient/request.rb:766:in >>>>>>>>>> `transmit' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/rest-client-2.0.0/lib/restclient/request.rb:215:in >>>>>>>>>> `execute' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/rest-client-2.0.0/lib/restclient/request.rb:52:in >>>>>>>>>> `execute' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/rest-client-2.0.0/lib/restclient/resource.rb:51:in >>>>>>>>>> `get' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/kubeclient-1.1.4/lib/kubeclient/common.rb:328:in >>>>>>>>>> `block in api' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/kubeclient-1.1.4/lib/kubeclient/common.rb:58:in >>>>>>>>>> `handle_exception' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/kubeclient-1.1.4/lib/kubeclient/common.rb:327:in >>>>>>>>>> `api' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/kubeclient-1.1.4/lib/kubeclient/common.rb:322:in >>>>>>>>>> `api_valid?' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluent-plugin-kubernetes_metadata_fil >>>>>>>>>> ter-0.24.0/lib/fluent/plugin/filter_kubernetes_metadata.rb:167:in >>>>>>>>>> `configure' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/agent.rb:144:in >>>>>>>>>> `add_filter' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/agent.rb:61:in >>>>>>>>>> `block in configure' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/agent.rb:57:in >>>>>>>>>> `each' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/agent.rb:57:in >>>>>>>>>> `configure' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/root_agent.rb:83:in >>>>>>>>>> `block in configure' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/root_agent.rb:83:in >>>>>>>>>> `each' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/root_agent.rb:83:in >>>>>>>>>> `configure' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/engine.rb:129:in >>>>>>>>>> `configure' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/engine.rb:103:in >>>>>>>>>> `run_configure' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/supervisor.rb:483:in >>>>>>>>>> `run_configure' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/supervisor.rb:154:in >>>>>>>>>> `block in start' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/supervisor.rb:360:in >>>>>>>>>> `call' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/supervisor.rb:360:in >>>>>>>>>> `main_process' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/supervisor.rb:333:in >>>>>>>>>> `block in supervise' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/supervisor.rb:332:in >>>>>>>>>> `fork' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/supervisor.rb:332:in >>>>>>>>>> `supervise' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/supervisor.rb:150:in >>>>>>>>>> `start' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/lib/fluent/command/fluentd.rb:173:in >>>>>>>>>> `<top (required)>' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/rubygems/rubygems/core_ext/kernel_require.rb:55:in >>>>>>>>>> `require' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /usr/share/rubygems/rubygems/core_ext/kernel_require.rb:55:in >>>>>>>>>> `require' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/gems/fluentd-0.12.23/bin/fluentd:5:in `<top >>>>>>>>>> (required)>' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/bin/fluentd:23:in `load' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: >>>>>>>>>> /opt/app-root/src/bin/fluentd:23:in `<main>' >>>>>>>>>> 2016-08-15 03:45:18 -0400 [error]: fluentd main process died >>>>>>>>>> unexpectedly. restarting. >>>>>>>>>> >>>>>>>>>> Other fluentd pods are fine: >>>>>>>>>> >>>>>>>>>> 2016-08-15 02:55:58 -0400 [info]: reading config file >>>>>>>>>> path="/etc/fluent/fluent.conf" >>>>>>>>>> >>>>>>>>>> However, I'm not getting any logs on pods running on nodes that >>>>>>>>>> have fluentd deployed properly. >>>>>>>>>> >>>>>>>>>> Any thoughts will be much appreciated :) >>>>>>>>>> >>>>>>>>>> Frank >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> users@lists.openshift.redhat.com >>>>>>>>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
_______________________________________________ users mailing list users@lists.openshift.redhat.com http://lists.openshift.redhat.com/openshiftmm/listinfo/users