I think the attachment was stripped off from the mail :( .
here's a public link.

https://drive.google.com/file/d/0B_z8xmsby0uxRDZEeWpLcnR2b3M/view?usp=sharing

On 25 September 2015 at 09:59, Lorenzo Fundaró <
lorenzo.fund...@dawandamail.com> wrote:

> This is the last logs i've got, even with a higher zkClientTimeout of 30s.
>
> this is a replica:
>
> 9/24/2015, 8:14:46 PMWARNRecoveryStrategyStopping recovery for
> core=dawanda coreNodeName=core_node69/24/2015, 8:14:56 PMWARN
> RecoveryStrategyStopping recovery for core=dawanda 
> coreNodeName=core_node69/24/2015,
> 8:16:34 PMWARNRecoveryStrategyStopping recovery for core=dawanda
> coreNodeName=core_node69/24/2015, 8:16:37 PMWARNPeerSyncPeerSync:
> core=dawanda url=http://solr6.dawanda.services:8983/solr too many updates
> received since start - startingUpdates no longer overlaps with our
> currentUpdates9/24/2015, 8:16:40 PMERRORReplicationHandlerSnapPull failed
> :org.apache.solr.common.SolrException: Index fetch failed :9/24/2015,
> 8:16:40 PMERRORRecoveryStrategyError while trying to
> recover:org.apache.solr.common.SolrException: Replication for recovery
> failed.9/24/2015, 8:16:40 PMERRORRecoveryStrategyRecovery failed - trying
> again... (0) core=dawanda9/24/2015, 8:16:55 
> PMERRORSolrCorenull:org.apache.lucene.store.AlreadyClosedException:
> Already closed:
> MMapIndexInput(path="/srv/loveos/solr/server/solr/dawanda/data/index.20150924174642739/_ulvt_Lucene50_0.tim")9/24/2015,
> 8:16:59 PMWARNSolrCore[dawanda] PERFORMANCE WARNING: Overlapping
> onDeckSearchers=29/24/2015, 8:17:53 PMWARNUpdateLogStarting log replay
> tlog{file=/srv/loveos/solr/server/solr/dawanda/data/tlog/tlog.0000000000000024343
> refcount=2} active=true starting pos=33935659/24/2015, 8:18:47 PMWARN
> RecoveryStrategyStopping recovery for core=dawanda coreNodeName=core_node6
> 9/24/2015, 8:18:57 PMWARNUpdateLogLog replay finished.
> recoveryInfo=RecoveryInfo{adds=2556 deletes=2455 deleteByQuery=0 errors=0
> positionOfStart=3393565}9/24/2015, 8:18:57 PMWARNRecoveryStrategyStopping
> recovery for core=dawanda coreNodeName=core_node69/24/2015, 8:19:07 PMWARN
> RecoveryStrategyStopping recovery for core=dawanda 
> coreNodeName=core_node69/24/2015,
> 8:19:17 PMWARNRecoveryStrategyStopping recovery for core=dawanda
> coreNodeName=core_node69/24/2015, 8:19:27 PMWARNRecoveryStrategyStopping
> recovery for core=dawanda coreNodeName=core_node69/24/2015, 8:19:37 PMWARN
> RecoveryStrategyStopping recovery for core=dawanda 
> coreNodeName=core_node69/24/2015,
> 8:19:47 PMWARNRecoveryStrategyStopping recovery for core=dawanda
> coreNodeName=core_node6
>
> and *again an overlapping searcher*.
>
> this is the leader:
>
> 9/24/2015, 8:18:56 PMWARNDistributedUpdateProcessorError sending update
> to http://solr6.dawanda.services:8983/solr9/24/2015, 8:18:56 PMERROR
> StreamingSolrClientserror9/24/2015, 8:18:56 PMWARN
> DistributedUpdateProcessorError sending update to
> http://solr6.dawanda.services:8983/solr9/24/2015, 8:18:56 PMWARN
> ZkControllerLeader is publishing core=dawanda coreNodeName =core_node6
> state=down on behalf of un-reachable replica
> http://solr6.dawanda.services:8983/solr/dawanda/; forcePublishState? 
> false9/24/2015,
> 8:18:56 PMERRORDistributedUpdateProcessorSetting up to try to start
> recovery on replica http://solr6.dawanda.services:8983/solr/dawanda/ after:
> org.apache.http.conn.ConnectionPoolTimeoutException: Timeout waiting for
> connection from pool9/24/2015, 8:18:57 
> PMERRORStreamingSolrClientserror9/24/2015,
> 8:18:57 PMWARNDistributedUpdateProcessorError sending update to
> http://solr6.dawanda.services:8983/solr9/24/2015, 8:18:57 PMWARN
> ZkControllerLeader is publishing core=dawanda coreNodeName =core_node6
> state=down on behalf of un-reachable replica
> http://solr6.dawanda.services:8983/solr/dawanda/; forcePublishState? 
> false9/24/2015,
> 8:18:57 PMERRORDistributedUpdateProcessorSetting up to try to start
> recovery on replica http://solr6.dawanda.services:8983/solr/dawanda/ after:
> org.apache.http.conn.ConnectionPoolTimeoutException: Timeout waiting for
> connection from pool
>
>
> I am doing jmx monitoring and right when two replicas went into recovering
> and down respectively, I checked the threads on the leader and there were
> at least 30 updatesExecutor running.
>
> I did a check on one of the replicas gc logs, and with gcviewer i get this
> for
>
> -* the leader*: http://snag.gy/UZyBu.jpg  , http://snag.gy/azrqH.jpg ,
> http://snag.gy/vef8D.jpg , http://snag.gy/mRVF0.jpg (on the cpu usage of
> this one, the graph going up and down from 20% to 90% cpu usage lasts for
> time that what is shown on the image, around 15 to 20 min)
> - *one replica: *http://snag.gy/PmW4t.jpg , http://snag.gy/2fpUK.jpg ,
> http://snag.gy/1Z5S2.jpg , http://snag.gy/AdlZw.jpg ,
> http://snag.gy/W4Kko.jpg ,
>
> As shown on the article about updates you sent me, when the time of the
> incident yesterday (one three replicas went down and only the leader was
> up) I saw a sustained spike on cpu,
> but on the contrary to what the article says, the replicas where having
> almost no cpu activity, mainly because they were down.
>
> I attached a thread dump of the leader and one replica. Again, I saw way
> too many updatesExecutors either in "park" or in "running" mode. On the
> replica I saw ~3000 starting with qtp, e.g.: "qtp1042863993-1487".
>
> I think what I am facing here is a sudden spike of (massive?) updates
> coming from our clients. And I think (need to recheck, but almost sure)
> that the updates request are not being done in batches. Maybe that could
> explain the massive amount of updatesExecutors, which assuming from their
> names, are the ones responsible for handling document updates.
>
> I will try to get more information about the clients and act on upon that.
>
> Thank you !
>
>
> On 23 September 2015 at 19:40, Lorenzo Fundaró <
> lorenzo.fund...@dawandamail.com> wrote:
>
>> here are the logs that didnt make it through the image: (sorry for the
>> misalignment on the logs)
>>
>> 9/23/2015, 7:14:49 PMERRORStreamingSolrClientserror
>>
>> org.apache.http.conn.ConnectionPoolTimeoutException: Timeout waiting for 
>> connection from pool
>>      at 
>> org.apache.http.impl.conn.PoolingClientConnectionManager.leaseConnection(PoolingClientConnectionManager.java:226)
>>      at 
>> org.apache.http.impl.conn.PoolingClientConnectionManager$1.getConnection(PoolingClientConnectionManager.java:195)
>>      at 
>> org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:422)
>>      at 
>> org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:863)
>>      at 
>> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82)
>>      at 
>> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:106)
>>      at 
>> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:57)
>>      at 
>> org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.run(ConcurrentUpdateSolrClient.java:233)
>>      at 
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>      at 
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>      at java.lang.Thread.run(Thread.java:745)
>>
>> 9/23/2015, 7:14:49 PMWARNDistributedUpdateProcessorError sending update
>> to http://solr8.dawanda.services:8983/solr
>>
>> org.apache.http.conn.ConnectionPoolTimeoutException: Timeout waiting for 
>> connection from pool
>>      at 
>> org.apache.http.impl.conn.PoolingClientConnectionManager.leaseConnection(PoolingClientConnectionManager.java:226)
>>      at 
>> org.apache.http.impl.conn.PoolingClientConnectionManager$1.getConnection(PoolingClientConnectionManager.java:195)
>>      at 
>> org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:422)
>>      at 
>> org.apache.http.impl.client.AbstractHttpClient.doExecute(AbstractHttpClient.java:863)
>>      at 
>> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82)
>>      at 
>> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:106)
>>      at 
>> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:57)
>>      at 
>> org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrClient$Runner.run(ConcurrentUpdateSolrClient.java:233)
>>      at 
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>      at 
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>      at java.lang.Thread.run(Thread.java:745)
>>
>> 9/23/2015, 7:14:49 PMWARNZkControllerLeader is publishing core=dawanda
>> coreNodeName =core_node10 state=down on behalf of un-reachable replica
>> http://solr8.dawanda.services:8983/solr/dawanda/; forcePublishState? 
>> false9/23/2015,
>> 7:14:49 PMERRORDistributedUpdateProcessorSetting up to try to start
>> recovery on replica http://solr8.dawanda.services:8983/solr/dawanda/ after:
>> org.apache.http.conn.ConnectionPoolTimeoutException: Timeout waiting for
>> connection from pool9/23/2015, 7:14:49 
>> PMERRORStreamingSolrClientserror9/23/2015,
>> 7:14:49 PMWARNDistributedUpdateProcessorError sending update to
>> http://solr8.dawanda.services:8983/solr
>>
>>
>> I am pretty sure I saw the overlapping searches but I would forget that
>> for now, since it hasn't appeared again. Maybe it was an old log that I
>> confused.
>>
>> I am gonna check the GC logs, and get back with some information.
>>
>> From the logs, I think the obvious action to take would be to increment
>> the zkClientTimeout to 30s (currently 15s)..
>>
>> On 23 September 2015 at 18:43, Erick Erickson <erickerick...@gmail.com>
>> wrote:
>>
>>> bq: and when one of the nodes enter recovery mode the leader says this:
>>>
>>> Hmmm, nothing came through, the mail filter is pretty aggressive
>>> about stripping attachments though.
>>>
>>> bq: You mean 10 seconds apart ?
>>>
>>> Hmmm, no I mean 10 minutes. That would
>>> explain the overlapping searchers since the only
>>> time you should be opening a new searcher is
>>> when you do a soft commit. But from the log
>>> fragment you pasted, that isn't happening so
>>> I'm totally mystified why you were getting those.
>>> Or was that some time ago and things have
>>> changed since then? I totally don't
>>> understand why you'd be seeing overlapping
>>> searchers, so grasping at straws. Given
>>> what you've added, I suspect this is totally
>>> off base.
>>>
>>> I see you're printing GC stats, are any of those
>>> GC times exceeding 15 seconds for stop-the-world
>>> pauses? That can trigger ZK marking the node as
>>> down and the node going into recovery. But with
>>> 16G heaps, that is unlikely, but you can see in
>>> your logs. If you are getting long stop-the-world
>>> pauses, perhaps you're very close to your JVM limit?
>>> In _some_ cases I've seen adding more memory
>>> decrease GC thrashing, the scenario is this:
>>>
>>> You're running very close to the limit of your JVM
>>> and each collection only reclaims a very little bit of
>>> memory (say 0.05%), so it goes immediately into
>>> another GC cycle. This is totally a shot in the dark
>>> though.
>>>
>>> Anyway, it's mystifying so far, but from everything
>>> you've written your system should not be going into
>>> recovery and you shouldn't be seeing
>>> overlapping searchers, so I'm missing something
>>>
>>> Best,
>>> Erick
>>>
>>> On Wed, Sep 23, 2015 at 9:25 AM, Lorenzo Fundaró <
>>> lorenzo.fund...@dawandamail.com> wrote:
>>>
>>> > On 23 September 2015 at 18:08, Erick Erickson <erickerick...@gmail.com
>>> >
>>> > wrote:
>>> >
>>> > > Wow, this is not expected at all. There's no
>>> > > way you should, on the face of it, get
>>> > > overlapping on-deck searchers.
>>> > >
>>> > > I recommend you put your maxWarmingSearchers
>>> > > back to 2, that's a fail-safe that is there to make
>>> > > people look at why they're warming a bunch of
>>> > > searchers at once.
>>> > >
>>> >
>>> > Ok, will do.
>>> >
>>> >
>>> > >
>>> > > With those settings, it's saying that autowarming is
>>> > > taking over 10 minutes.
>>> >
>>> >
>>> > What do you mean ? I don't think the autowarming is taking 10 minutes
>>> here.
>>> > Do you maybe mean 10 seconds apart ?
>>> > Every 10 min I issue a soft commit which would create a new searcher
>>> > with several autowarmings, one for each cache. Looking at the stats the
>>> > slowest autowarm is from filter cache and it takes around 2 seconds.
>>> >
>>> >
>>> > > This is absurdly long, so either
>>> > > something is pathologically wrong with your Solr
>>> > > or you're really committing more often than you think.
>>> > > Possibly you have a client issuing commits?
>>> >
>>> >
>>> > mm no, I checked my client and I am not doing any commit.
>>> > Here are some logs from the leader:
>>> >
>>> > INFO  - 2015-09-23 16:21:59.803;
>>> > org.apache.solr.update.DirectUpdateHandler2; start
>>> >
>>> >
>>> commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
>>> > INFO  - 2015-09-23 16:22:00.981;
>>> org.apache.solr.core.SolrDeletionPolicy;
>>> > SolrDeletionPolicy.onCommit: commits: num=2
>>> > commit{dir=NRTCachingDirectory(MMapDirectory@
>>> > /srv/loveos/solr/server/solr/dawanda/data/index.20150818133228229
>>> > lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@66cd8efb
>>> ;
>>> > maxCacheMB=48.0
>>> maxMergeSizeMB=4.0),segFN=segments_ilp,generation=24109}
>>> > commit{dir=NRTCachingDirectory(MMapDirectory@
>>> > /srv/loveos/solr/server/solr/dawanda/data/index.20150818133228229
>>> > lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@66cd8efb
>>> ;
>>> > maxCacheMB=48.0
>>> maxMergeSizeMB=4.0),segFN=segments_ilq,generation=24110}
>>> > INFO  - 2015-09-23 16:22:00.981;
>>> org.apache.solr.core.SolrDeletionPolicy;
>>> > newest commit generation = 24110
>>> > INFO  - 2015-09-23 16:22:01.010;
>>> > org.apache.solr.update.DirectUpdateHandler2; end_commit_flush
>>> > INFO  - 2015-09-23 16:22:16.967;
>>> > org.apache.solr.update.DirectUpdateHandler2; start
>>> >
>>> >
>>> commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
>>> > INFO  - 2015-09-23 16:22:18.452;
>>> org.apache.solr.core.SolrDeletionPolicy;
>>> > SolrDeletionPolicy.onCommit: commits: num=2
>>> > commit{dir=NRTCachingDirectory(MMapDirectory@
>>> > /srv/loveos/solr/server/solr/dawanda/data/index.20150818133228229
>>> > lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@66cd8efb
>>> ;
>>> > maxCacheMB=48.0
>>> maxMergeSizeMB=4.0),segFN=segments_ilq,generation=24110}
>>> >
>>> > You can look at your Solr logs and see commits, just
>>> > > look for the word "commit". When reading those lines,
>>> > > it'll say whether it has openSearcher true or false.
>>> > > Are the timestamps when openSearcer=true really
>>> > > 10 minutes apart?
>>> > >
>>> >
>>> > You mean 10 seconds apart ?
>>> >
>>> >
>>> > >
>>> > > You'll also see autowarm times in your logs, see how
>>> > > long they really take. If they really take 10 minutes,
>>> > > we need to get to the bottom of that because the
>>> > > autowarm counts you're showing in your cache
>>> > > configurations don't indicate any problem here.
>>> > >
>>> > > Bottom line:
>>> > > 1> you shouldn't be seeing nodes go into recovery in the
>>> > > first place. Are your Solr logs showing any ERROR
>>> > > level messages?
>>> > >
>>> > > 2> it's extremely surprising that you're getting any
>>> > > overlapping on-deck searchers. If it turns out that
>>> > > your autowarming is really taking more than a few
>>> > > seconds, getting a stack trace to see where Solr is
>>> > > spending all the time is warranted.
>>> > >
>>> >
>>> > Will do.
>>> >
>>> > >
>>> > > 3> Any clues from the logs _why_ they're going
>>> > > into recovery? Also look at your leader's log file
>>> > > and see if there are any messages about "leader
>>> > > initiated recovery". If you see that, then perhaps
>>> > > one of the timeouts is too short.
>>> > >
>>> > > 4> the tlog size is quite reasonable. It's only relevant
>>> > > when a node goes down for some reason anyway,
>>> > > so I wouldn't expend too much energy worrying about
>>> > > them until we get to the bottom of overlapping
>>> > > searchers and nodes going into recovery.
>>> > >
>>> > > BTW, nice job of laying out the relevant issues and
>>> > > adding supporting information! I wish more problem
>>> > > statements were as complete. If your Solr is 4.7.0,
>>> > > there was a memory problem and you should definitely
>>> > > go to 4.7.2. The symptom here is that you'll see
>>> > > Out of Memory errors...
>>> > >
>>> > >
>>> > > Best,
>>> > > Erick
>>> >
>>> >
>>> > Thank you very much !
>>> >
>>> > >
>>> >
>>> >
>>> > > On Wed, Sep 23, 2015 at 8:48 AM, Lorenzo Fundaró <
>>> > > lorenzo.fund...@dawandamail.com> wrote:
>>> > >
>>> > > > Hi !,
>>> > > >
>>> > > > I keep getting nodes that fall into recovery mode and then issue
>>> the
>>> > > > following log WARN every 10 seconds:
>>> > > >
>>> > > > WARN   Stopping recovery for core=xxxx coreNodeName=core_node7
>>> > > >
>>> > > > and sometimes this appears as well:
>>> > > > PERFORMANCE WARNING: Overlapping onDeckSearchers=2
>>> > > > At higher traffic time, this gets worse and out of 4 nodes only 1
>>> is
>>> > up.
>>> > > > I have 4 solr nodes each running two cores A and B of 13GB and
>>> 1.5GB
>>> > > > respectively. Core A gets a lot of index updates and higher query
>>> > traffic
>>> > > > compared to core B. Core A is going through active/recovery/down
>>> states
>>> > > > very often.
>>> > > > Nodes are coordinated via Zookeeper, we have three, running in
>>> > different
>>> > > > machines than Solr.
>>> > > > Each machine has around 24 cores and between 38 and 48 GB of RAM,
>>> with
>>> > > each
>>> > > > Solr getting 16GB of heap memory.
>>> > > > I read this article:
>>> > > >
>>> > > >
>>> > >
>>> >
>>> https://lucidworks.com/blog/understanding-transaction-logs-softcommit-and-commit-in-sorlcloud/
>>> > > >
>>> > > > and decided to apply:
>>> > > >
>>> > > >      <autoCommit>
>>> > > >        <!-- Every 15 seconds -->
>>> > > >        <maxTime>${solr.autoCommit.maxTime:15000}</maxTime>
>>> > > >        <openSearcher>false</openSearcher>
>>> > > >      </autoCommit>
>>> > > >
>>> > > > and
>>> > > >
>>> > > >      <autoSoftCommit>
>>> > > >        <!-- Every 10 minutes -->
>>> > > >        <maxTime>${solr.autoSoftCommit.maxTime:600000}</maxTime>
>>> > > >      </autoSoftCommit>
>>> > > >
>>> > > > I also have these cache configurations:
>>> > > >
>>> > > >     <filterCache class="solr.LFUCache"
>>> > > >                  size="64"
>>> > > >                  initialSize="64"
>>> > > >                  autowarmCount="32"/>
>>> > > >
>>> > > >     <queryResultCache class="solr.LRUCache"
>>> > > >                      size="512"
>>> > > >                      initialSize="512"
>>> > > >                      autowarmCount="0"/>
>>> > > >
>>> > > >     <documentCache class="solr.LRUCache"
>>> > > >                    size="1024"
>>> > > >                    initialSize="1024"
>>> > > >                    autowarmCount="0"/>
>>> > > >
>>> > > >     <cache name="perSegFilter"
>>> > > >       class="solr.search.LRUCache"
>>> > > >       size="10"
>>> > > >       initialSize="0"
>>> > > >       autowarmCount="10"
>>> > > >       regenerator="solr.NoOpRegenerator" />
>>> > > >
>>> > > >        <fieldValueCache class="solr.FastLRUCache"
>>> > > >                         size="512"
>>> > > >                         autowarmCount="0"
>>> > > >                         showItems="32" />
>>> > > >
>>> > > > I also have this:
>>> > > > <maxWarmingSearchers>6</maxWarmingSearchers>
>>> > > > The size of the tlogs are usually between 1MB to 8MB.
>>> > > > I thought the changes above could improve the situation, but I am
>>> not
>>> > > 100%
>>> > > > convinced they did since after 15 min one of the nodes entered
>>> recovery
>>> > > > mode again.
>>> > > >
>>> > > > any ideas ?
>>> > > >
>>> > > > Thanks in advance.
>>> > > >
>>> > > > Cheers !
>>> > > >
>>> > > > --
>>> > > >
>>> > > > --
>>> > > > Lorenzo Fundaro
>>> > > > Backend Engineer
>>> > > > E-Mail: lorenzo.fund...@dawandamail.com
>>> > > >
>>> > > > Fax       + 49 - (0)30 - 25 76 08 52
>>> > > > Tel        + 49 - (0)179 - 51 10 982
>>> > > >
>>> > > > DaWanda GmbH
>>> > > > Windscheidstraße 18
>>> > > > 10627 Berlin
>>> > > >
>>> > > > Geschäftsführer: Claudia Helming, Michael Pütz
>>> > > > Amtsgericht Charlottenburg HRB 104695 B
>>> > > >
>>> > >
>>> >
>>> >
>>> >
>>> > --
>>> >
>>> > --
>>> > Lorenzo Fundaro
>>> > Backend Engineer
>>> > E-Mail: lorenzo.fund...@dawandamail.com
>>> >
>>> > Fax       + 49 - (0)30 - 25 76 08 52
>>> > Tel        + 49 - (0)179 - 51 10 982
>>> >
>>> > DaWanda GmbH
>>> > Windscheidstraße 18
>>> > 10627 Berlin
>>> >
>>> > Geschäftsführer: Claudia Helming, Michael Pütz
>>> > Amtsgericht Charlottenburg HRB 104695 B
>>> >
>>>
>>
>>
>>
>> --
>>
>> --
>> Lorenzo Fundaro
>> Backend Engineer
>> E-Mail: lorenzo.fund...@dawandamail.com
>>
>> Fax       + 49 - (0)30 - 25 76 08 52
>> Tel        + 49 - (0)179 - 51 10 982
>>
>> DaWanda GmbH
>> Windscheidstraße 18
>> 10627 Berlin
>>
>> Geschäftsführer: Claudia Helming, Michael Pütz
>> Amtsgericht Charlottenburg HRB 104695 B
>>
>
>
>
> --
>
> --
> Lorenzo Fundaro
> Backend Engineer
> E-Mail: lorenzo.fund...@dawandamail.com
>
> Fax       + 49 - (0)30 - 25 76 08 52
> Tel        + 49 - (0)179 - 51 10 982
>
> DaWanda GmbH
> Windscheidstraße 18
> 10627 Berlin
>
> Geschäftsführer: Claudia Helming, Michael Pütz
> Amtsgericht Charlottenburg HRB 104695 B
>



-- 

-- 
Lorenzo Fundaro
Backend Engineer
E-Mail: lorenzo.fund...@dawandamail.com

Fax       + 49 - (0)30 - 25 76 08 52
Tel        + 49 - (0)179 - 51 10 982

DaWanda GmbH
Windscheidstraße 18
10627 Berlin

Geschäftsführer: Claudia Helming, Michael Pütz
Amtsgericht Charlottenburg HRB 104695 B

Reply via email to