: Is it possible to sort the facet pivot items by using a stat value like sum
: or mean ?
Not at the moment no.
It's something i'd like to see implemented eventually, but it would
require a *huge* amount of re-implementation of how the faceting code
currently works.
-Hoss
what does your schema.xml look like?
what does your solrconfig.xml look like?
what does the document you are indexing look like?
what is the full error with stack trace from your server logs?
details matter.
https://wiki.apache.org/solr/UsingMailingLists
: Date: Thu, 12 Mar 2015 20:27:05
1) the error message you posted doesn't appear to have been copied
verbatim (note: ...or dex...) ... please proivde the *exact* error you
are getting -- ideally with full stack trace from the solr logs.
2) the field/ you posted doens't match the field name in your error
message
3)providing a
: is a syntactically significant character to the query parser, so it's
getting confused by it in the text of your query.
you're seeing the same problem as if you tried to search for foo:bar in
the yak field using q=yak:foo:bar
you either need to backslash escape the : characters, or wrap the
: The TOC I meant was as in a traditional print publication with section #s,
I believe supressing the printed page TOC was a deliberate choice made in
the PDF stylesheet ... IIRC it was about 10 (printed) pages long and
didn't seem to add any value since hte PDF metadata seemed much more
https://people.apache.org/~hossman/#threadhijack
Thread Hijacking on Mailing Lists
When starting a new discussion on a mailing list, please do not reply to
an existing message, instead start a fresh email. Even if you change the
subject line of your email, other mail headers still track which
: Here's my need : I'd like to log Solr Responses so as to achieve some
: business statistics.
: I'd like to report, as a daily/weekly/yearly/whateverly basis, the following
: KPIs :
...
: I think I'll soon get into performance issues, as you guess.
: Do you know a better approach ?
All of
: So are we still using the jetty? Are we still dependent on war file?
As explained in the ref guide...
https://cwiki.apache.org/confluence/display/solr/Major+Changes+from+Solr+4+to+Solr+5#MajorChangesfromSolr4toSolr5-SolrisNowaStandaloneServer
Internally, Solr is still implemented via Servlet
: There were no changes made in the solrconfig.xml file except added that
: updateRequestProcessorChain default=true/updateRequestProcessorChain
: block.
ok, first off: if you already *had* another updateRequestProcessorChain
that said 'default=true' just adding a new one would be weird and
: In 5.0 the new way is to not use a servlet container, just use the
: start/stop scripts.
More specifically...
https://cwiki.apache.org/confluence/display/solr/Major+Changes+from+Solr+4+to+Solr+5
Internally, Solr is still implemented via Servlet APIs and is powered by
Jetty -- but this is
: If your expire_at_dt field is not populated automatically, let's step
: back and recheck a sanity setting. You said it is a managed schema? Is
: it a schemaless as well? With an explicit processor chain? If that's
: the case, your default chain may not be running AT ALL.
yeah ... my only
: Following query posts a document and sets expire_at_dt explicitly. That
: is working perfectly ok and ducument expires at defined time.
so the delete trigge logic is working correctly...
: But when trying to post with TTL (following query), document does not
: expire after given time.
: Unfortunately, this is all 5.1 and instructs me to run the 'start from
: scratch' process.
a) checkout the left nav of any ref guide page webpage which has a link to
Older Versions of this Guide (PDF)
b) i'm not entirely sure i understand what you're asking, but i'm guessing
you mean...
*
: We are using Solr. We would not configure two different Solr instances to
: write to the same index. So why would a normal Solr set-up possibly end
: up having more than one process writing to the same index?
The risk here is that if you configure lockType=single, and then have some
: Hi! Solr is starting up dormant for me, until a client wake it up with a
: REST request, or I open admin UI, only then the remaining initializing
: happens.
: Is it something known?
based on my recollection of the servlet spec, that sounds like a
bug/glitch/config option in your Servlet
: But when I use SolrJ and get it as object, I am seeing that the UTC date is
: of type Date and I am not able to retrieve back the UTC date from it and I
: get only long timestamp from that object.
:
: I also see a private variable in that Date class called as cDate which has
: what I want
: to get the UTC back which I thought might not be required as already the
: cDate field in that Date class is having the UTC date.
general suggestion: your life will be a lot easier if you stop looking at
the implementation details of JVM classes -- just because your current JVM
implements
: I need to uniquely identify a document inside of a Similarity class during
: scoring. Is it possible to get value of unique key of a document at this
: point?
Can you tell us a bit more about your usecase ... your problem description
is a bit vague, and sounds like it may be an XY Problem...
: 1. name:DocumentOne^7 = doc1(score=7)
: 2. name:DocumentOne^7 AND place:notExist^3 = doc1(score=7)
: 3. place:(34\ High\ Street)^3 = doc1(score=3), doc2(score=3)
: 4. name:DocumentOne^7 OR place:(34\ High\ Street)^3 = doc1(score=10),
: doc2(score=3)
...
: it's not clear why you need
: Sure, sorry I did not do it before, I just wanted to take minimum of your
: valuable time. So in my custom Similarity class I am trying to implement such
: a logic, where score calculation is only based on field weight and a field
: match - that's it. In other words, if a field matches the
: Hi,
:
: You are using /update when registering, but using /update/extract when
invoking.
:
: Ahmet
if your goal is that *every* doc will get a last_modified, regarldess of
how it is indexed, then you don't need to set the update.chain default
on every requestHandler -- instead just mark
: 1. Look further down in the stack trace for the caused by that details
: the specific cause of the exception.
: I am still not able to find the cause of this.
jack is refering to the log file from your server ... sometimes there
are more details there.
: Sorry i but don't know it is
: I'm trying to start a SolrCloud cluster with a kerberized Zookeeper. I'm not
: sure if it is possible, I have a Hadoop Cluster with an already running
: zookeeper and I do not think running two zoo in parallel would be the wise
: choice.
: Is there a way to use SASL with SolrCloud ?
Work has
: The collection fails to be created (shard_replica dir and data and index
: across the servers get created but collection creation fails)
:
: The full log is appended below. I thought it should be a straight forward
: class not found problem but I just can't seem to fix this (few hours now).
:
: I think until Solr become completely standalone, it could be major task for
Solr 5.0 is already completley standalone.
Running bin/solr (or bin/solr.cmd) as a standalone daemon is the only
documented supported way to run Solr 5.
*Internally* Solr is using jetty -- but that is 100% an
: processor
: class=solr.processor.DocExpirationUpdateProcessorFactory
: int name=autoDeletePeriodSeconds30/int
: str name=ttlFieldNamettl/str
: str name=expirationFieldNameexpire_at/str
: /processor
...
: And I have included the
:
: Thanks for giving multiple options , I ll try them out both ,but last time
: I checked, having +60SECONDS as the default value for ttl was giving me
: an invalid date format exception, I am assuming that would only be the
that's because ttl should not be a date field -- it should be a
: We are trying to do a POC for searching our log files with a single node
Solr(396 GB RAM with 14 TB Space).
: Since the server is powerful, added 2 Billion records successfully search
is working fine without much issues.
how much CPU?
Assuming it's comparable to the amount of RAM you've
: Recently, we have switched over to use atomic update instead of re-indexing
: when we need to update a doc in the index. It looks to me that the
: timestamp field is not updated during an atomic update. I have also looked
: into TimestampUpdateProcessorFactory and it looks to me that won't
: of official documentation, but I wonder abstractly how a non-committer then
: should contribute to the documentation. I just did an evaluation of
...
: With current technology, possibilities include:
you pretty much nailed it...
: * Make a comment within Confluence suggesting
: I had a discussion with @search_mb about this on IRC, and he explained how
: my collection query would still work with user6, though we couldn't
: resolve why the solr Core Admin page doesn't show the collection name as
: user6.
Core Admin pages in the UI are still specific to *CORES* ... no
: Because they have different potential authors, the two systems now serve
: different purposes.
:
: There are still some pages on the MoinMoin wiki that contain
: documentation that should be in the reference guide, but isn't.
:
: The MoinMoin wiki is still useful, as a place where users can
: We upgraded our cluster to Solr 4.10.0 for couple days and again
: reverted back to 4.8.0. However the dashboard still shows Solr 4.10.0.
: Do you know why?
because you didn't fully revert - you are still running Solr 4.10.0 - the
details of what steps you took to try and switch back make a
: But my concern is if solr internally creates a different set for each
: and every different queries upon sort values and they lasts for ever I
: think.
https://cwiki.apache.org/confluence/display/solr/Pagination+of+Results
Cursors in Solr are a logical concept, that doesn't involve
(cross posted, please confine replies to general@lucene)
ApacheCon 2015 Will be in Austin Texas April 13-17.
http://apachecon.com/
The Call For Papers is currently open, but it ends 2015-02-01 (11:55PM GMT-0600)
: However, you will notice on page 228, under the section Suggester, it
: gives an example of a suggester search component using
: solr.SpellCheckComponet.
...
: So it would appear the solr.SuggestComponent has been around since 4.7,
: but the documentation has not caught up with the
: If you just need to transform an individual result, that can be done by a
: custom DocTransformer. But from your email, I think you need a custom
: SearchComponent.
if your PostFilter has already collected all of the info you need, and you
now just wnat to return a subset of that information
and see what error
you get.
:
: Darren
:
: On Wed, Jan 21, 2015 at 6:48 PM, Chris Hostetter hossman_luc...@fucit.org
: wrote:
:
:
: : I posted a question on stackoverflow but in hindsight this would have
: been
: : a better place to start. Below is the link.
: :
: : Basically I can't get
1) which version of Solr are you using? (note that the online HTML ref
guide is a DRARFT that applies to 5.0 - you may want to review the
specific released version of the ref guide that applies to your version of
solr: http://archive.apache.org/dist/lucene/solr/ref-guide/
2) the behavior of
On Wed, 21 Jan 2015, Mihran Shahinian wrote:
: Date: Wed, 21 Jan 2015 16:06:18 -0600
: From: Mihran Shahinian slowmih...@gmail.com
: Reply-To: solr-user@lucene.apache.org
: To: solr-user@lucene.apache.org
: Subject: boosting by geodist - GC Overhead Limit exceeded
:
: I am running solr 4.10.2
: I'm facing a problem with multiple field sort in Solr. I'm using the
: following fields in sort :
:
: PublishDate asc,DocumentType asc
correction: you are using: PublishDate desc,DocumentType desc
: The sort is only happening on PublishDate, DocumentType seemsto completely
: ignored.
: I posted a question on stackoverflow but in hindsight this would have been
: a better place to start. Below is the link.
:
: Basically I can't get the example working when using an external ZK cluster
: and auto-core discovery. Solr 4.10.1 works fine, but the newest release
your SO URL shows
to
make it easy for people. I remember it being a struggle for me when I was a
new user, and I know we get lots of questions on the user-list about it.
:
: My apologies to you for not answering this sooner.
:
: James Dyer
: Ingram Content Group
:
:
: -Original Message-
: From: Chris
: As a foolish dev (not malicious I hope!), I did mess around with something
: like this once; I was writing my own Codec. I found I had to create a file
: called META-INF/services/org.apache.lucene.codecs.Codec in my solr plugin jar
: that contained the fully-qualified class name of my codec: I
: Does anybody know for sure whether the stats component fully supports
: distributed mode? It is listed in the doc as supporting distributed mode
it's been supported for as long as i can remember -- since Day 1 of the
StatsComponent i believe.
: (at least for old, non-SolrCloud distrib mode),
: assuming I've written the subclass of the postings format, I need to tell
: Solr to use it.
:
: Do I just do something like:
:
: fieldType name=ocr class=solr.TextField postingsFormat=MySubclass /
the postingFormat xml tag in schema.xml just refers to the name of the
postingFormat in SPI --
: ...the nuts bolts of it is that the PostingFormat baseclass should take
: care of all the SPI name registration that you need based on what you
: pass to the super() construction ... allthough now that i think about it,
: i'm not sure how you'd go about specifying your own name for the
:
: This is starting to sound pretty complicated. Are you saying this is not
: doable with Solr 4.10?
it should be doable in 4.10, using a wrapper class like the one i
mentioned below (delegating to Lucene51PostingsFormat instead of
Lucene50PostingsFormat) ... it's just that the 4.10 APIs are
:I am stuck at a strange issue, I have my custom Query Component that has
: to load spring application context for some additional runtime filtering of
: records.
:I have included my jars as dependency in solrConfig.xml, SOLR is able to
: load my plugin but spring appplication fails to
: It looks like this is a good starting point:
:
: http://wiki.apache.org/solr/SolrConfigXml#codecFactory
The default SchemaCodecFactory already supports defining a diff posting
format per fieldType - but there isn't much in solr to let you tweak
individual options on specific posting formats
: Thank you for your reply Chris :) Solr is producing the correct result on
: its own. The problem is that I am calling a dataload class to call Solr,
: which worked for assigned ID and composite ID, but not for UUID. Is there a
Sorry -- still confused: are you confirming that you've tracked
i don't have specific answers toall of your questions, but you should
probably look at SOLR-445 where a lot of this has already ben discussed
and multiple patches with different approaches have been started...
https://issues.apache.org/jira/browse/SOLR-445
: Date: Wed, 7 Jan 2015 12:38:47
: However the facets I am getting for the date is till last month, say today
: is 24th December and I am getting it till 24th November. How should I
: modify my query to obtain results till today? Tried a few options using HIT
: and TRIAL :) but could not arrive at a solution.
it's not clear
: It's a single Solr Instance, and in my files, I used 'doc_key' everywhere,
: but I changed it to id in the email I sent out wanting to make it easier
: to read, sorry don't mean to confuse you :)
https://wiki.apache.org/solr/UsingMailingLists
- what version of solr?
- how exactly are you
: I am exploring faceting in SOLR in collection1 example Faceting fields are
: defined in solrconfig.xml under browse request handler which is used in
: in-built VelocityResponseWriter
context is everything -- you cut out the key line that would answer
explain your question...
: #1 is a trade off against being possibly more available to writes in the case
: of a single down node. In the cloud case, you're still open for business. In
: the classical replication case, you're no longer available for writes if the
: downed node is the master.
or to put it another way:
: I'm trying to get anything to index. Starting with the simplest file
: possible. As it stands no extraction is working. I'm just trying to get any
: extraction working. I've followed that guide, I'll try again.
let's back up for a minute.
You have a plain text file, and you want to index it.
This sounds a lot like...
https://issues.apache.org/jira/browse/SOLR-6643
: Date: Fri, 12 Dec 2014 16:54:03 -0700 (MST)
: From: solr-user solr-u...@hotmail.com
: Reply-To: solr-user@lucene.apache.org
: To: solr-user@lucene.apache.org
: Subject: Re: Solr 4.10.2 Found core but I get No cores
: When I run the following query (Solr 4.10.2) with edit-distance, I'm
: getting a null pointer exception:
:
: *host/solr/select?q=fld:(Event
: ID)fl=strdist(eventid,fld_alphaonly,edit)*
probably this bug: https://issues.apache.org/jira/browse/SOLR-6540
: responselst name=errorstr
: But how do we avoid this error?
: Is there anyway to tweak the query and return empty result instead of null
: pointer exception?
did you look at the issue i linked to?
: probably this bug: https://issues.apache.org/jira/browse/SOLR-6540
A workarround in some contexts can be to wrap the
Ok, so i've been working on updating hte ref guide to account for hte new
way to run the examples in 5.0.
The spell checking page...
https://cwiki.apache.org/confluence/display/solr/Spell+Checking
...has some examples that loosely corroloate to the techproducts
example, but even if
(NOTE: cross posted to several lucene lists, if you have replies, please
confine them to general@lucene)
-- Forwarded message --
In case you've missed it:
- ApacheCon North America returns to Austin, Texas, 13-17 April 2015
http://apachecon.com/
- Call for Papers open
sounds like this bug...
https://issues.apache.org/jira/browse/SOLR-6154
...in which case it has nothing to do with your use of multiple
collections, it's just dependent on wether or not the first node to
respond happens to have a doc in every range bucket .. any bucket
missing (because of
: ?I have a SolrCloud cluster with two servers and I created a collection using
two shards with this command:
...
: There were 230 documents in the set I indexed and there were 3 different
prefixes (RM!, WW! and BH!) but all were routed into the same shard. Is there
anything I can do
: No, I wasn't aware of these. I will give that a try. If I stop the Solr
: jetty service manually, things recover fine, but the hang occurs when I
: 'stop' or 'terminate' the EC2 instance. The Zookeeper leader reports a
I don't know squat about AWS Auto-Scaling, (and barely anything about AWS)
: can you please include the *exact* solrconfig.xml schema.xml you are
: using for coreA ... you've given us an overview of what you changed, but
: that's not enough for anyone to actally try and reproduce your problem.
if it helps (since hte list doesn't allow attachments) feel free to open a
: coreA (which has our modified schema and solrconfig) is found by solr but is
: not shown in the solr admin page:
can you please include the *exact* solrconfig.xml schema.xml you are
using for coreA ... you've given us an overview of what you changed, but
that's not enough for anyone to
: With an implicit collection, is it legal to index the same document
: (same unique ID) in 2 different shards? I know, it kind of defeats the
: purpose of having a unique ID...
Each doc (defined by uniqueKey) must exist in one and only one shard ...
when this constraint is violated, you'll
: We are trying out solr 4.10.2 (as an upgrade from 4.0) and are seeing an odd
: issue.
...
: Note that we are testing under Windows 7, and that the sample solr in 4.10.2
: runs fine (with the same folder structure, etc tho with the default schema
: and solrconfig.xml). I have run thru
in general, a facet count of 0 means the term is in the index but does not
match an of the docs in the result set.
if you are doing a query that matches all docs, and seeing facet values
with a mincount of 0, that means the *term* is still in the index, but the
documents that contained those
: But I am curious about other peoples' experiences with SO. Do you ask
: questions in that forum? Do you answer? Why? How do you compare that
: support channel with this one? Did you migrate from one to another?
: Private replies are welcome, though I suspect this topic might be
: interesting
: Subject: Re: Disappearance of post.jar from the new tutorial
:
: I removed reference to it as the same class is in solr-core's JAR.
:
: The idea is to hide the details behind bin/post and before end of year
: (before 5.0 release at least) to get that taken care of.
This doesn't make any
: Solr is not really designed to be extended in this way. In fact I believe
: they are moving towards an architecture where this is even less possible -
Correct.
Starting with 5.0, the fact that servlets a servlet container are used
by solr becomes a pure implementation detail - subject to
: Subject: REST API Alternative to admin/luke
this smells like an XY problem ... if /admin/luke gives you the data you
want, why not use /admin/luke ? ... what is it about /admin/luke that
prevents you from solving your problem? what is your ultimate goal?
: If I use the admin/luke URL all
: I did not oversee a feature of the rest endpoints. So probably we will
: stick with the admin/luke endpoint to achieve our goal.
Ok ... i mean ... yeah -- the /admin/luke endpoint exists to tell you what
fields are *actually* in your index, regardless of who/how they are in
your index.
the
: I thought that the auto commit is per update handler because they are
: configured within the update handler tag.
updateHandler is not the same thing as a requestHandler that does
updates.
there can be many Update request handlers configured, but there is only
ever one updateHandler/ in a
: For sorting DocValues are the best option I think.
yep, definitely a good idea.
: I have a usecase for using cursorpage and when tried to check this, I got
: outOfMemory just for sorting by id.
what does the field/fieldType for your uniqueKey field look like?
If you aren't using
: An oversight I think. If you create a patch, let me know and we can
: get it committed.
that definitely sounds bad we should certainly try to fix that before 5.0
comes out since it does have back-compat implictations...
https://issues.apache.org/jira/browse/SOLR-6718
...better to
I am not a security expert, but in my opinion the safest way to run solr
securely is to forget all about usernames passwords and instead use
SSL with client SSL certificates...
https://cwiki.apache.org/confluence/display/solr/Enabling+SSL
: Date: Tue, 4 Nov 2014 12:53:30 +
: From: Shay
: I am currently working on SolrCloud and its related security
: configurations for securing Solr web applications using HTTP Basic
: Authentication mechanism. Among the Solr nodes inside the SolrCloud
: clustered env, there seem to be existing some inter-solr-node
: communication issues due
: Sure thing, but how do I get the results output in CSV format?
: response.getResults() is a list of SolrDocuments.
Either use something like the NoOpResponseParser which will give you the
entire response back as a single string, or implement your own
ResponseParser along hte lines of...
: Why do you want to use CSV in SolrJ? Alexandre are you looking for a
It's a legitmate question - part of providing good community support is
making sure we understand *why* users are asking how to do something, so
we can give good advice on other solutions people might not even have
Solr has never really owrked well with years prior to 1 because the
specs for how they should be formated/parsed -- in particular realted to
year 0 have always been painfully ambiguious/contradictory.
https://issues.apache.org/jira/browse/SOLR-2773
If you are really trying to deal with year 0
: I was just trying to index the fields returned by my msql and i found this
If you are importing dates from MySql where you have -00-00T00:00:00Z
as the default value, you should actaully be getting an error lsat time i
checked, but this explains the right way to tell the MySQL JDBC
: fl=id field(units_used) archive_id
I didn't even realize until today that fl was documented to support space
seperated fields. i've only ever used commas...
fl=id,field(units_used),archive_id
Please go ahead and file a bug in Jira for this, and note in the summary
that using commas
: I am getting below error while doing ant dist .
The build system (up to 4.10.1) was unintentinally requiring that javadoc
jars existed -- and this recently manifested as a problem when this
particular javadoc jar somehow fvanished from maven.org.
This issue tracks the fix which will be n
: I tried using this function
: boost=recip(ms(NOW/HOUR,startdatez,3.16e-11.0,0.08,0.05))
: but it fails with this error:
: org.apache.lucene.queryparser.classic.ParseException: Expected ')' at
: position 29 in 'recip(ms(NOW/HOUR,startdatez,3.16e-11.0,0.08,0.05))'
look very carefully at your
: I meant signature will be broken. For example suppose the destination of
: hash function for signature fields are sig. After each partial update it
: becomes: 00!
details please.
how are you configuring your update processor chain? what does your schema
look like? what types of
: I am attempting to create a script (java script) using the
: StatelessScriptUpdateProcessorFactory feature of solr but I am blocked
: on how to access the current core instance (ultimately to access it's
: schema)? In the wikipedia example the input document is accessible using
: doc =
: It's for an optimization. If the keyword is 'match all docs', I want to
: remove a custom PostFilter from the query and change the sort parameters
: (so the app doesn't have to do it). It looks like the responseHeader is
: displaying the 'originalParams', which are immutable.
that is in fact
you can still use the SignatureUpdateProcessorFactory for your usecase,
just don't configure teh signatureField to be the same as your uniqueKey
field.
configure some othe fieldname (ie signature) instead.
: Date: Tue, 14 Oct 2014 12:08:26 +0330
: From: Ali Nazemian alinazem...@gmail.com
:
You're seeing an excentric edge case of the way purely negative Boolean
Queries work.
in this example...
description:ruby AND (-title:java)
you have a top level boolean query with two mandatory clauses. the second
clause in that top level query is a (bested) boolean query with a
: Since we are moving to Java 8, how about we support Nashorn?
:
: http://winterbe.com/posts/2014/04/05/java8-nashorn-tutorial/
define support ? what exactly do you have in mind ?
The two places i can think of in solr that support scripting already
support Nashorn by default (assuming it's
: I am trying to index a collection that has child documents. I have
: successfully loaded the data into my index using SolrJ, and I have
: verified that I can search correctly using the child of method in my
: fq variable. Now, I would like to use Velocity (Solritas) to display
: the
: i try to add documents to the index and boost them (hole document) but i
: get this error message:
:
: ERROR org.apache.solr.core.SolrCore –
: org.apache.solr.common.SolrException: Error parsing JSON field value.
: Unexpected OBJECT_START
:
: Any ideas?
The top level structure you are
: Subject: Having an issue with pivot faceting
Ok - first off -- your example request doens't include any facet.pivot
params, so you aren't using pivot faceting .. which makes me concerned
that if you aren't using the feature you think you are, or don't
understand the feature you are using.
: Solr for large and time consuming queries. We have found a very inconsistent
: result in the time elapsed when pinging Solr. If we ping Solr from a desktop
: Windows 7 machine, there is usually a 5 ms elapsed time. But if we ping the
: same Solr instance from a Windows Server 2008 machine, it
: +1 for using a different cache, but that's being quite unfamiliar with the
: code.
in (a) common case, people tend to drill down and filter on facet
constraints -- so using a special purpose cache for the refinements would
result in redundent caching of the same info in multiple places.
:
: For eg : red shirt under 20$ should be translated to q=shirtfq=price:[*
: TO 20] and possibly apply color to one the attribute of doc index.
:
: in parser overrided method, how can i add the filter and pass the query
: back?
I don't think you can acomplish this just within the QParser API ...
: I am indexing Solr 4.9.0 using the /update request handler and am getting
: errors from Tika - Illegal IOException from
: org.apache.tika.parser.xml.DcXMLParser@74ce3bea which is caused by
: MalFormedByteSequenceException: Invalid byte 1 of 1-byte UTF-8 sequence. I
FWIW: that error appears to
401 - 500 of 4495 matches
Mail list logo