This is an automated email from the ASF dual-hosted git repository. reschke pushed a commit to branch OAK-11650 in repository https://gitbox.apache.org/repos/asf/jackrabbit-oak.git
commit 07cd90124d61dc76d721d79b15ee895b6ad12d7a Author: Julian Reschke <[email protected]> AuthorDate: Thu Apr 10 09:51:08 2025 +0100 OAK-11650: Fix Javadoc errors - one-line only </li> --- .../plugins/migration/version/VersionableEditor.java | 1 - .../jackrabbit/oak/plugins/version/VersionHook.java | 5 ----- .../readonly/ReadOnlyAuthorizationConfiguration.java | 1 - .../apache/jackrabbit/api/JackrabbitValueFactory.java | 4 ---- .../apache/jackrabbit/api/binary/BinaryUpload.java | 7 ------- .../codecs/lucene40/Lucene40DocValuesFormat.java | 1 - .../codecs/lucene40/Lucene40FieldInfosFormat.java | 2 -- .../codecs/lucene40/Lucene40PostingsFormat.java | 1 - .../codecs/lucene41/Lucene41PostingsFormat.java | 6 ------ .../codecs/lucene42/Lucene42FieldInfosFormat.java | 2 -- .../codecs/lucene46/Lucene46FieldInfosFormat.java | 2 -- .../lucene/search/similarities/IBSimilarity.java | 3 --- .../lucene/search/similarities/TFIDFSimilarity.java | 19 ------------------- 13 files changed, 54 deletions(-) diff --git a/oak-core/src/main/java/org/apache/jackrabbit/oak/plugins/migration/version/VersionableEditor.java b/oak-core/src/main/java/org/apache/jackrabbit/oak/plugins/migration/version/VersionableEditor.java index c39ec939c4..80bd35ce3f 100644 --- a/oak-core/src/main/java/org/apache/jackrabbit/oak/plugins/migration/version/VersionableEditor.java +++ b/oak-core/src/main/java/org/apache/jackrabbit/oak/plugins/migration/version/VersionableEditor.java @@ -52,7 +52,6 @@ import static org.apache.jackrabbit.oak.plugins.migration.version.VersionHistory * it can skip copying version histories and remove the * {@code mix:versionable} mixin together with any related * properties (see {@link VersionHistoryUtil#removeVersionProperties(NodeBuilder, TypePredicate)}). - * </li> * </ul> */ public class VersionableEditor extends DefaultEditor { diff --git a/oak-core/src/main/java/org/apache/jackrabbit/oak/plugins/version/VersionHook.java b/oak-core/src/main/java/org/apache/jackrabbit/oak/plugins/version/VersionHook.java index 00efacacd2..be42d7ec91 100644 --- a/oak-core/src/main/java/org/apache/jackrabbit/oak/plugins/version/VersionHook.java +++ b/oak-core/src/main/java/org/apache/jackrabbit/oak/plugins/version/VersionHook.java @@ -45,21 +45,16 @@ import static org.apache.jackrabbit.oak.spi.commit.CompositeHook.compose; * {@link VersionEditor} - creates version history, handles * checking-in, checking-out and restoring, prevents a * checked-in node from being modified, - * </li> * <li> * {@link VersionStorageEditor} - validates changes on the * version storage, - * </li> * </ul> - * </li> * <li> * {@link VersionableCollector} - collects all existing versionable * UUIDs, so assigned histories won't be removed in the next step, - * </li> * <li> * {@link OrphanedVersionCleaner} - removes all histories that are * empty and have no longer a parent versionable node. - * </li> * </ol> * */ diff --git a/oak-exercise/src/main/java/org/apache/jackrabbit/oak/exercise/security/authorization/models/readonly/ReadOnlyAuthorizationConfiguration.java b/oak-exercise/src/main/java/org/apache/jackrabbit/oak/exercise/security/authorization/models/readonly/ReadOnlyAuthorizationConfiguration.java index 6ca2624e35..d7be75517d 100644 --- a/oak-exercise/src/main/java/org/apache/jackrabbit/oak/exercise/security/authorization/models/readonly/ReadOnlyAuthorizationConfiguration.java +++ b/oak-exercise/src/main/java/org/apache/jackrabbit/oak/exercise/security/authorization/models/readonly/ReadOnlyAuthorizationConfiguration.java @@ -150,7 +150,6 @@ import static org.apache.jackrabbit.oak.spi.security.RegistrationConstants.OAK_S * to the list of required service IDs</li> * <li>make sure the 'Authorization Composition Type' is set to AND</li> * </ul> - * </li> * <li>Wait for the {@link org.apache.jackrabbit.oak.spi.security.SecurityProvider} to be successfully registered again.</li> * </ul> * diff --git a/oak-jackrabbit-api/src/main/java/org/apache/jackrabbit/api/JackrabbitValueFactory.java b/oak-jackrabbit-api/src/main/java/org/apache/jackrabbit/api/JackrabbitValueFactory.java index 0360fb037d..596a841b27 100644 --- a/oak-jackrabbit-api/src/main/java/org/apache/jackrabbit/api/JackrabbitValueFactory.java +++ b/oak-jackrabbit-api/src/main/java/org/apache/jackrabbit/api/JackrabbitValueFactory.java @@ -46,7 +46,6 @@ import org.osgi.annotation.versioning.ProviderType; * <li> * Direct Binary Access - enable a client to upload or download binaries * directly to/from a storage location - * </li> * </ul> * * <p> @@ -97,14 +96,12 @@ import org.osgi.annotation.versioning.ProviderType; * optionally choose to provide a {@link BinaryUploadOptions} via * {@link #initiateBinaryUpload(long, int, BinaryUploadOptions)} if * additional options must be specified. - * </li> * <li> * <b>Upload</b>: The remote client performs the actual binary upload * directly to the binary storage provider. The {@link BinaryUpload} * returned from the previous call to {@link * #initiateBinaryUpload(long, int)} contains detailed instructions on * how to complete the upload successfully. - * </li> * <li> * <b>Complete</b>: The remote client notifies the Jackrabbit-based * application that step 2 is complete. The upload token returned in @@ -114,7 +111,6 @@ import org.osgi.annotation.versioning.ProviderType; * with a regular {@link Binary JCR Binary} that can then be used to * write JCR content including the binary (such as an nt:file structure) * and persist it using {@link Session#save}. - * </li> * </ol> */ @ProviderType diff --git a/oak-jackrabbit-api/src/main/java/org/apache/jackrabbit/api/binary/BinaryUpload.java b/oak-jackrabbit-api/src/main/java/org/apache/jackrabbit/api/binary/BinaryUpload.java index 90181828bd..9ed50fa90b 100644 --- a/oak-jackrabbit-api/src/main/java/org/apache/jackrabbit/api/binary/BinaryUpload.java +++ b/oak-jackrabbit-api/src/main/java/org/apache/jackrabbit/api/binary/BinaryUpload.java @@ -80,7 +80,6 @@ import org.osgi.annotation.versioning.ProviderType; * If {@code (fileSize / maxPartSize) > numUploadURIs}, then the client * cannot proceed and will have to request a new set of URIs with the * right fileSize as {@code maxSize}. - * </li> * <li> * Calculate the {@code partSize} and the number of URIs to use. * <br> @@ -103,7 +102,6 @@ import org.osgi.annotation.versioning.ProviderType; * provided upload URI to upload the entire binary, with * {@code partSize = fileSize}. Note that it is not required to * use all of the URIs provided in {@code uploadURIs}. - * </li> * <li> * If {@code fileSize / partSize == numUploadURIs}, all part * URIs must to be used. The {@code partSize} to use for all @@ -112,7 +110,6 @@ import org.osgi.annotation.versioning.ProviderType; * It is also possible to simply use {@code maxPartSize} as the * value for {@code partSize} in this case, for every part * except the last. - * </li> * </ol> * Optionally, a client may select a different {@code partSize}, * for example if the client has more information about the @@ -126,16 +123,13 @@ import org.osgi.annotation.versioning.ProviderType; * (unless {@code maxPartSize = -1} meaning unlimited)</li> * <li>{@code partSize > (fileSize / numUploadURIs)}</li> * </ol> - * </li> * <li> * Upload: segment the binary into {@code partSize}, for each segment * take the next URI from {@code uploadURIs} (strictly in order), * proceed with a standard HTTP PUT for each, and for the last part use * whatever segment size is left. - * </li> * <li> * If a segment fails during upload, retry (up to a certain timeout). - * </li> * <li> * After the upload has finished successfully, notify the application, * for example through a complete request, passing the {@link @@ -157,7 +151,6 @@ import org.osgi.annotation.versioning.ProviderType; * Instead, simply restart the upload from the beginning by calling * {@link JackrabbitValueFactory#initiateBinaryUpload(long, int)} when * the situation preventing a successful upload has been resolved. - * </li> * </ol> * * <h2>Example JSON view</h2> diff --git a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40DocValuesFormat.java b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40DocValuesFormat.java index 02a98cf0fc..dcaf06f5b3 100644 --- a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40DocValuesFormat.java +++ b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40DocValuesFormat.java @@ -123,7 +123,6 @@ import org.apache.lucene.util.packed.PackedInts; * <li>{@code BYTES_VAR_STRAIGHT BYTES_VAR_STRAIGHT} in contrast to other straight * variants uses a <tt>.idx</tt> file to improve lookup perfromance. In contrast to * {@code BYTES_VAR_DEREF BYTES_VAR_DEREF} it doesn't apply deduplication of the document values. - * </li> * </ul> * <p> * Limitations: diff --git a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40FieldInfosFormat.java b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40FieldInfosFormat.java index e2c9c76763..bdb391f630 100644 --- a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40FieldInfosFormat.java +++ b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40FieldInfosFormat.java @@ -73,7 +73,6 @@ import org.apache.lucene.store.DataOutput; // javadoc * <li>If the eighth lowest-order bit is set (0x80), positions are omitted for the * indexed field.</li> * </ul> - * </li> * <li>DocValuesBits: a byte containing per-document value types. The type * recorded as two four-bit integers, with the high-order bits representing * <code>norms</code> options, and the low-order bits representing @@ -94,7 +93,6 @@ import org.apache.lucene.store.DataOutput; // javadoc * <li>12: fixed-length sorted byte array values. ({@code Type#BYTES_FIXED_SORTED BYTES_FIXED_SORTED})</li> * <li>13: variable-length sorted byte array values. ({@code Type#BYTES_VAR_SORTED BYTES_VAR_SORTED})</li> * </ul> - * </li> * <li>Attributes: a key-value map of codec-private attributes.</li> * </ul> * diff --git a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40PostingsFormat.java b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40PostingsFormat.java index e389d5448e..e52667a35d 100644 --- a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40PostingsFormat.java +++ b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene40/Lucene40PostingsFormat.java @@ -81,7 +81,6 @@ import org.apache.lucene.util.fst.FST; // javadocs * {@link DocsEnum#advance(int)}. Larger values result in smaller indexes, greater * acceleration, but fewer accelerable cases, while smaller values result in bigger indexes, * less acceleration (in case of a small value for MaxSkipLevels) and more accelerable cases. - * </li> * <li>MaxSkipLevels is the max. number of skip levels stored for each term in the .frq file. A * low value results in smaller indexes but less acceleration, a larger value results in * slightly larger indexes but greater acceleration. See format of .frq file for more diff --git a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene41/Lucene41PostingsFormat.java b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene41/Lucene41PostingsFormat.java index e88d0ee223..365bec13b7 100644 --- a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene41/Lucene41PostingsFormat.java +++ b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene41/Lucene41PostingsFormat.java @@ -62,7 +62,6 @@ import org.apache.lucene.util.packed.PackedInts; * that are all the same value are encoded in an optimized way.</p> * <p>In VInt blocks, integers are encoded as {@link DataOutput#writeVInt VInt}: * the block size is variable.</p> - * </li> * * <li> * <b>Block structure</b>: @@ -77,14 +76,12 @@ import org.apache.lucene.util.packed.PackedInts; * <position, payload length>, * <position, offset start, offset length>, and * <position, payload length, offsetstart, offset length>.</p> - * </li> * * <li> * <b>Skipdata settings</b>: * <p>The structure of skip table is quite similar to previous version of Lucene. Skip interval is the * same as block size, and each skip entry points to the beginning of each block. However, for * the first block, skip data is omitted.</p> - * </li> * * <li> * <b>Positions, Payloads, and Offsets</b>: @@ -102,7 +99,6 @@ import org.apache.lucene.util.packed.PackedInts; * <p>With this strategy, the majority of payload and offset data will be outside .pos file. * So for queries that require only position data, running on a full index with payloads and offsets, * this reduces disk pre-fetches.</p> - * </li> * </ul> * * <p> @@ -222,7 +218,6 @@ import org.apache.lucene.util.packed.PackedInts; * separately encode as packed blocks.</li> * </ol> * If frequencies are not omitted, PackedFreqBlock will be generated without d-gap step. - * </li> * <li>VIntBlock stores remaining d-gaps (along with frequencies when possible) with a format * that encodes DocDelta and Freq: * <p>DocDelta: if frequencies are indexed, this determines both the document @@ -239,7 +234,6 @@ import org.apache.lucene.util.packed.PackedInts; * <p>If frequencies were omitted ({@link IndexOptions#DOCS_ONLY}) it would be this * sequence of VInts instead:</p> * <p>7,4</p> - * </li> * <li>PackedDocBlockNum is the number of packed blocks for current term's docids or frequencies. * In particular, PackedDocBlockNum = floor(DocFreq/PackedBlockSize) </li> * <li>TrimmedDocFreq = DocFreq % PackedBlockSize == 0 ? DocFreq - 1 : DocFreq. diff --git a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene42/Lucene42FieldInfosFormat.java b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene42/Lucene42FieldInfosFormat.java index e3f00e9e38..1affb2ea9f 100644 --- a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene42/Lucene42FieldInfosFormat.java +++ b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene42/Lucene42FieldInfosFormat.java @@ -74,7 +74,6 @@ import org.apache.lucene.store.DataOutput; // javadoc * <li>If the eighth lowest-order bit is set (0x80), positions are omitted for the * indexed field.</li> * </ul> - * </li> * <li>DocValuesBits: a byte containing per-document value types. The type * recorded as two four-bit integers, with the high-order bits representing * <code>norms</code> options, and the low-order bits representing @@ -85,7 +84,6 @@ import org.apache.lucene.store.DataOutput; // javadoc * <li>2: BinaryDocValues. ({@code DocValuesType#BINARY})</li> * <li>3: SortedDocValues. ({@code DocValuesType#SORTED})</li> * </ul> - * </li> * <li>Attributes: a key-value map of codec-private attributes.</li> * </ul> * diff --git a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene46/Lucene46FieldInfosFormat.java b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene46/Lucene46FieldInfosFormat.java index ce4cdbcf00..57bbf53f96 100755 --- a/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene46/Lucene46FieldInfosFormat.java +++ b/oak-lucene/src/main/java/org/apache/lucene/codecs/lucene46/Lucene46FieldInfosFormat.java @@ -76,7 +76,6 @@ import org.apache.lucene.store.DataOutput; * <li>If the eighth lowest-order bit is set (0x80), positions are omitted for the * indexed field.</li> * </ul> - * </li> * <li>DocValuesBits: a byte containing per-document value types. The type * recorded as two four-bit integers, with the high-order bits representing * <code>norms</code> options, and the low-order bits representing @@ -87,7 +86,6 @@ import org.apache.lucene.store.DataOutput; * <li>2: BinaryDocValues. ({@code DocValuesType#BINARY})</li> * <li>3: SortedDocValues. ({@code DocValuesType#SORTED})</li> * </ul> - * </li> * <li>DocValuesGen is the generation count of the field's DocValues. If this is -1, * there are no DocValues updates to that field. Anything above zero means there * are updates stored by {@link DocValuesFormat}.</li> diff --git a/oak-lucene/src/main/java/org/apache/lucene/search/similarities/IBSimilarity.java b/oak-lucene/src/main/java/org/apache/lucene/search/similarities/IBSimilarity.java index 7a371966c9..9de818166d 100644 --- a/oak-lucene/src/main/java/org/apache/lucene/search/similarities/IBSimilarity.java +++ b/oak-lucene/src/main/java/org/apache/lucene/search/similarities/IBSimilarity.java @@ -56,7 +56,6 @@ import org.apache.lucene.search.similarities.Normalization.NoNormalization; * <li>{@link DistributionLL}: Log-logistic</li> * <li>{@link DistributionLL}: Smoothed power-law</li> * </ul> - * </li> * <li>{@link Lambda}: λ<sub>w</sub> parameter of the * probability distribution * <ul> @@ -65,11 +64,9 @@ import org.apache.lucene.search.similarities.Normalization.NoNormalization; * <li>{@link LambdaTTF}: <code>F<sub>w</sub>/N</code> or * average number of occurrences of w in the collection</li> * </ul> - * </li> * <li>{@link Normalization}: Term frequency normalization * <blockquote>Any supported DFR normalization (listed in * {@link DFRSimilarity})</blockquote> - * </li> * </ol> * <p> * @see DFRSimilarity diff --git a/oak-lucene/src/main/java/org/apache/lucene/search/similarities/TFIDFSimilarity.java b/oak-lucene/src/main/java/org/apache/lucene/search/similarities/TFIDFSimilarity.java index 959201ce10..7554eed06d 100644 --- a/oak-lucene/src/main/java/org/apache/lucene/search/similarities/TFIDFSimilarity.java +++ b/oak-lucene/src/main/java/org/apache/lucene/search/similarities/TFIDFSimilarity.java @@ -138,35 +138,29 @@ import org.apache.lucene.util.BytesRef; * To avoid this problem, a different document length normalization * factor is used, which normalizes to a vector equal to or larger * than the unit vector: <i>doc-len-norm(d)</i>. - * </li> * * <li>At indexing, users can specify that certain documents are more * important than others, by assigning a document boost. * For this, the score of each document is also multiplied by its boost value * <i>doc-boost(d)</i>. - * </li> * * <li>Lucene is field based, hence each query term applies to a single * field, document length normalization is by the length of the certain field, * and in addition to document boost there are also document fields boosts. - * </li> * * <li>The same field can be added to a document during indexing several times, * and so the boost of that field is the multiplication of the boosts of * the separate additions (or parts) of that field within the document. - * </li> * * <li>At search time users can specify boosts to each query, sub-query, and * each query term, hence the contribution of a query term to the score of * a document is multiplied by the boost of that query term <i>query-boost(q)</i>. - * </li> * * <li>A document may match a multi term query without containing all * the terms of that query (this is correct for some of the queries), * and users can further reward documents matching more query terms * through a coordination factor, which is usually larger when * more terms are matched: <i>coord-factor(q,d)</i>. - * </li> * </ul> * * <p>Under the simplifying assumption of a single field in the index, @@ -218,7 +212,6 @@ import org.apache.lucene.util.BytesRef; * <ul> * <li><i>Query-boost</i> for the query (actually for each query term) * is known when search starts. - * </li> * * <li>Query Euclidean norm <i>|V(q)|</i> can be computed when search starts, * as it is independent of the document being scored. @@ -242,14 +235,11 @@ import org.apache.lucene.util.BytesRef; * There are other applications that may require this. * And this is exactly what normalizing the query vector <i>V(q)</i> * provides: comparability (to a certain extent) of two or more queries. - * </li> * * <li>Applying query normalization on the scores helps to keep the * scores around the unit vector, hence preventing loss of score data * because of floating point precision limitations. - * </li> * </ul> - * </li> * * <li>Document length norm <i>doc-len-norm(d)</i> and document * boost <i>doc-boost(d)</i> are known at indexing time. @@ -257,7 +247,6 @@ import org.apache.lucene.util.BytesRef; * is saved as a single value in the index: <i>norm(d)</i>. * (In the equations below, <i>norm(t in d)</i> means <i>norm(field(t) in doc d)</i> * where <i>field(t)</i> is the field associated with term <i>t</i>.) - * </li> * </ul> * * <p><i>Lucene's Practical Scoring Function</i> is derived from the above. @@ -329,7 +318,6 @@ import org.apache.lucene.util.BytesRef; * </tr> * </table> * <br> <br> - * </li> * * <li> * <A NAME="formula_idf"></A> @@ -364,7 +352,6 @@ import org.apache.lucene.util.BytesRef; * </tr> * </table> * <br> <br> - * </li> * * <li> * <A NAME="formula_coord"></A> @@ -376,7 +363,6 @@ import org.apache.lucene.util.BytesRef; * {@link #coord(int, int) coord(q,d)} * by the Similarity in effect at search time. * <br> <br> - * </li> * * <li><b> * <A NAME="formula_queryNorm"></A> @@ -442,7 +428,6 @@ import org.apache.lucene.util.BytesRef; * </table> * <br> <br> * - * </li> * * <li> * <A NAME="formula_termBoost"></A> @@ -458,7 +443,6 @@ import org.apache.lucene.util.BytesRef; * and so the boost of a term in the query is accessible by calling the sub-query * {@link org.apache.lucene.search.Query#getBoost() getBoost()}. * <br> <br> - * </li> * * <li> * <A NAME="formula_norm"></A> @@ -468,12 +452,10 @@ import org.apache.lucene.util.BytesRef; * <li><b>Field boost</b> - set by calling * {@link org.apache.lucene.document.Field#setBoost(float) field.setBoost()} * before adding the field to a document. - * </li> * <li><b>lengthNorm</b> - computed * when the document is added to the index in accordance with the number of tokens * of this field in the document, so that shorter fields contribute more to the score. * LengthNorm is computed by the Similarity class in effect at indexing. - * </li> * </ul> * The {@link #computeNorm} method is responsible for * combining all of these factors into a single float. @@ -505,7 +487,6 @@ import org.apache.lucene.util.BytesRef; * </table> * Note that search time is too late to modify this <i>norm</i> part of scoring, * e.g. by using a different {@link Similarity} for search. - * </li> * </ol> * * @see org.apache.lucene.index.IndexWriterConfig#setSimilarity(Similarity)
