Hi Everton,

Thanks for the suggestions. I made the changes to the config files on both
nodes as you suggested. Since it is not possible for me to force the client
to do a source specific join I added the following line at interface ra_ap0
on node 2 where the client is attached:

interface ra_ap0
ip igmp
ip igmp query-interval 125
ip igmp query-max-response-time-dsec 100
ip igmp join 239.255.255.250 192.168.4.60

I do see the source-specific IGMPv3 join group 239.255.255.250 for source
192.168.4.60 which is addressed to 224.0.0.22 on the side of node2. However
this join request never makes it to node 1 where the source is located on
ra_ap0.
Shouldn't the pimd route this join request to the node where the source is
attached ?

Thanks,




On Mon, Oct 26, 2009 at 6:44 AM, Everton Marques
<[email protected]>wrote:

> Hi,
>
> You did not mention whether you got a source-specific IGMPv3 join to the
> channel (S,G)=(192.168.4.60,239.255.255.250). Please notice qpimd is
> unable to program the multicast forwarding cache with non-source-specific
> groups. Usually the key issue is to instruct the receiver application to
> join the source-specific channel (S,G).
>
> Regarding the config, the basic rule is:
> 1) Enable "ip pim ssm" everywhere (on every interface that should pass
> mcast).
> 2) Enable both "ip pim ssm" and "ip igmp" on interfaces attached to
> the receivers (IGMPv3 hosts).
>
> An even simpler config rule to remember is to enable both commands
> everywhere. They should not cause any harm.
>
> Hence, if your mcast receiver is attached to Node 2 at  ra_ap0, I think you
> will
> need at least the following config:
>
> !
> ! Node 1
> !
> interface ra_ap0
>  ip pim ssm
> interface ra_sta0
>  ip pim ssm
>
> !
> ! Node 2
> !
> interface ra_ap0
>  ip pim ssm
>  ip igmp
> interface ra_sta0
>  ip pim ssm
>
> Hope this helps,
> Everton
>
> On Mon, Oct 26, 2009 at 4:42 AM, Yoda geek <[email protected]>
> wrote:
> > Hi Everton & Fellow  qpimd users,
> >
> > We're trying to stream multicast video traffic between a Tversity server
> and
> > a multicast client separated by 2 nodes (node1 and node2). Each node is
> > running quagga suite (version 0.99.15) along with qpimd (version 0.158)
> > running on top of Linux 2.6.26.
> > Node 1 has 3 network interfaces - eth0, ap0 and ra_sta0
> > Node 2 has 2 network interfaces - ra_sta0 and ra_ap0
> > The Tversity server talks to interface ra_ap0 on Node 1 and the multicast
> > client talks to interface ra_ap0 on Node 2
> > Nodes 1 and 2 talk with each other over their ra_sta0 interfaces
> >
> > Below is a graphical depiction :
> >
> > Tversity server   -----------ra_ap0--> Node 1
> > --ra_sta0-----------------ra_sta0-->Node
> > 2-----ra_ap0------------------------> Video Client
> > ===========             ======================
> > ======================                      =============
> >
> >
> > Node 1 pimd.conf file
> > ==================
> > !
> > ! Zebra configuration saved from vty
> > ! 2009/08/01 20:26:06
> > !
> > hostname node1
> > password zebra
> > enable password zebra
> > log stdout
> > !
> > interface eth0
> > !
> > interface eth1
> > !
> > interface lo
> > !
> > interface ra_ap0
> > ip pim ssm
> > ip igmp
> > ip igmp query-interval 125
> > ip igmp query-max-response-time-dsec 100
> > ip igmp join 239.255.255.250 192.168.4.60
> > !
> > interface ra_sta0
> > ip igmp
> > ip igmp query-interval 125
> > ip igmp query-max-response-time-dsec 100
> > !
> > !
> > ip multicast-routing
> > !
> > line vty
> > !
> >
> > Node 2 pimd.conf configuration file
> > ============================
> > !
> > ! Zebra configuration saved from vty
> > ! 2009/08/02 21:54:14
> > !
> > hostname node2
> > password zebra
> > enable password zebra
> > log stdout
> > !
> > interface eth0
> > !
> > interface eth1
> > !
> > interface lo
> > !
> > interface ra_ap0
> > ip igmp
> > ip igmp query-interval 125
> > ip igmp query-max-response-time-dsec 100
> > ip igmp join 239.255.255.250 192.168.4.60
> > !
> > interface ra_sta0
> > ip igmp
> > ip igmp query-interval 125
> > ip igmp query-max-response-time-dsec 100
> > !
> > !
> > ip multicast-routing
> > !
> > line vty
> > !
> >
> > From the above configuration you can see that interface ra_ap0 on node 1
> is
> > configured to be multicast source (ip pim ssm).
> > We do see some multicast join requests in wireshark from both the server
> and
> > the client however no data flow. Initially we started qpimd without
> > the entry "igmp join ..." on either client side node or server side node.
> > Looking at node 1 configuration through "show  ip igmp groups" we didn't
> see
> > the group membership for "239.255.255.250" while this group membership
> was
> > observed on node 2. I put this group membership on both nodes to force
> them
> > to join this multicast group - however without success.
> >
> > Just to give you a background - when both client and server are talking
> to
> > same node - say node 2 and same interface ra_ap0 (without qpimd running)
> > multicast video gets served flawlessly from Tversity server to client
> > through the node.
> > But with the 2 node setup we aren't able to see the video streams go
> through
> > to the client.
> >
> > Could you please review  the above configuration for errors or have any
> > suggestions to reseolve this issue ? Any help would be greatly
> appreciated.
> >
> > Thanks,
> >
> >
>

Reply via email to