Hi Eunsu,

By going through the documentation, I think you are right, you shouldn't
use withUsedHostsPerRemoteDc because it will contact nodes in other
datacenters.  No i don't use withUsedHostsPerRemoteDc, but instead i use
withLocalDc option.

On Tue, Sep 18, 2018 at 11:02 AM, Eunsu Kim <eunsu.bil...@gmail.com> wrote:

> Yes, I altered the system_auth key space before adding the data center.
>
> However, I suspect that the new data center did not get the system_auth
> data and therefore could not authenticate to the client. Because the new
> data center did not get the replica count by altering keyspace.
>
> Do your clients have the 'withUsedHostsPerRemoteDc' option?
>
>
> On 18 Sep 2018, at 1:17 PM, Pradeep Chhetri <prad...@stashaway.com> wrote:
>
> Hello Eunsu,
>
> I am also using PasswordAuthenticator in my cassandra cluster. I didn't
> come across this issue while doing the exercise on preprod.
>
> Are you sure that you changed the configuration of system_auth keyspace
> before adding the new datacenter using this:
>
> ALTER KEYSPACE system_auth WITH REPLICATION = {'class':
> 'NetworkTopologyStrategy', 'datacenter1': '3'};
>
> Regards,
> Pradeep
>
>
>
> On Tue, Sep 18, 2018 at 7:23 AM, Eunsu Kim <eunsu.bil...@gmail.com> wrote:
>
>>
>> In my case, there were authentication issues when adding data centers.
>>
>> I was using a PasswordAuthenticator.
>>
>> As soon as the datacenter was added, the following authentication error
>> log was recorded on the client log file.
>>
>> com.datastax.driver.core.exceptions.AuthenticationException:
>> Authentication error on host /xxx.xxx.xxx.xx:9042: Provided username apm
>> and/or password are incorrect
>>
>> I was using DCAwareRoundRobinPolicy, but I guess it's probably because of
>> the withUsedHostsPerRemoteDc option.
>>
>> I took several steps and the error log disappeared. It is probably
>> ’nodetool rebuild' after altering the system_auth table.
>>
>> However, the procedure was not clearly defined.
>>
>>
>> On 18 Sep 2018, at 2:40 AM, Pradeep Chhetri <prad...@stashaway.com>
>> wrote:
>>
>> Hello Alain,
>>
>> Thank you very much for reviewing it. You answer on seed nodes cleared my
>> doubts. I will update it as per your suggestion.
>>
>> I have few followup questions on decommissioning of datacenter:
>>
>> - Do i need to run nodetool repair -full on each of the nodes (old + new
>> dc nodes) before starting the decommissioning process of old dc.
>> - We have around 15 apps using cassandra cluster. I want to make sure
>> that all queries before starting the new datacenter are going with right
>> consistency level i.e LOCAL_QUORUM instead of QUORUM. Is there a way i can
>> log the consistency level of each query somehow in some log file.
>>
>> Regards,
>> Pradeep
>>
>> On Mon, Sep 17, 2018 at 9:26 PM, Alain RODRIGUEZ <arodr...@gmail.com>
>> wrote:
>>
>>> Hello Pradeep,
>>>
>>> It looks good to me and it's a cool runbook for you to follow and for
>>> others to reuse.
>>>
>>> To make sure that cassandra nodes in one datacenter can see the nodes of
>>>> the other datacenter, add the seed node of the new datacenter in any of the
>>>> old datacenter’s nodes and restart that node.
>>>
>>>
>>> Nodes seeing each other from the distinct rack is not related to seeds.
>>> It's indeed recommended to use seeds from all the datacenter (a couple or
>>> 3). I guess it's to increase availability on seeds node and/or maybe to
>>> make sure local seeds are available.
>>>
>>> You can perfectly (and even have to) add your second datacenter nodes
>>> using seeds from the first data center. A bootstrapping node should never
>>> be in the list of seeds unless it's the first node of the cluster. Add
>>> nodes, then make them seeds.
>>>
>>>
>>> Le lun. 17 sept. 2018 à 11:25, Pradeep Chhetri <prad...@stashaway.com>
>>> a écrit :
>>>
>>>> Hello everyone,
>>>>
>>>> Can someone please help me in validating the steps i am following to
>>>> migrate cassandra snitch.
>>>>
>>>> Regards,
>>>> Pradeep
>>>>
>>>> On Wed, Sep 12, 2018 at 1:38 PM, Pradeep Chhetri <prad...@stashaway.com
>>>> > wrote:
>>>>
>>>>> Hello
>>>>>
>>>>> I am running cassandra 3.11.3 5-node cluster on AWS with SimpleSnitch.
>>>>> I was testing the process to migrate to GPFS using AWS region as the
>>>>> datacenter name and AWS zone as the rack name in my preprod environment 
>>>>> and
>>>>> was able to achieve it.
>>>>>
>>>>> But before decommissioning the older datacenter, I want to verify that
>>>>> the data in newer dc is in consistence with the one in older dc. Is there
>>>>> any easy way to do that.
>>>>>
>>>>> Do you suggest running a full repair before decommissioning the nodes
>>>>> of older datacenter ?
>>>>>
>>>>> I am using the steps documented here: https://medium.com/p/465
>>>>> e9bf28d99 I will be very happy if someone can confirm me that i am
>>>>> doing the right steps.
>>>>>
>>>>> Regards,
>>>>> Pradeep
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>
>>
>>
>
>

Reply via email to