Hi,

Which Apache Hop version are you using?
There used to be an issue with the description that ended up instead of the
name of the execution information location, but it was fixed quite a while
ago.




On Tue, Mar 25, 2025 at 2:32 PM AndersonCode42 <[email protected]>
wrote:

> Hi Bart, I did as you suggested (see attachments), and it worked :)
>
> And it is related to what Davide Cisco mentioned.
>
> So we have a bug in this auditing part, and we avoid using it for now?
>
> I think I will open another issue with these details and the workaround
> you described.
>
> Thanks a lot Bart and Davide
>
>
>
> Em ter., 25 de mar. de 2025 às 09:59, Bart Maertens <[email protected]>
> escreveu:
>
>> Hi,
>>
>> Lazy conversion is no longer available in the Table Input transform, but
>> it's still mentioned in the docs, so that's a bug in the docs that will be
>> fixed soon. Thanks for pointing that out!
>> The logs aren't complete, but It looks like the pipeline did execute
>> correctly. The problem seems to be in your pipeline run configuration. Can
>> you check your pipeline run configuration in the metadata perspective?
>> Can you clear out any values you have for "Execution Information
>> Location" and "Execution data profile", save your pipeline run
>> configuration and see if the error is still there when you run again?
>>
>> Bart
>>
>> On Tue, Mar 25, 2025 at 1:42 PM AndersonCode42 <[email protected]>
>> wrote:
>>
>>> Hello everyone, I'm a guy who is migrating from Pentaho Kettle to Apache
>>> Hop and I've already faced my first obstacle.
>>>
>>> *Background*:
>>> OS: Lubuntu 24.10
>>> Java: Temurin - 21.0.6+7 - LTS - JRE
>>> Apache Hop: apache-hop-client-2.12.0.zip
>>>
>>>
>>> I have a simple pipeline that basically should copy data from a table,
>>> then use "Select values" (*hop.SelectValues.jpg*) to convert the field
>>> types, and finally write it to another database table.
>>>
>>> However, when processing the rows, there is a row with a specific value
>>> that throws an error in the log (*logging.txt*) and stops processing
>>> the remaining rows. Strange fact: despite interrupting the processing, the
>>> pipeline visually appears to have run successfully, since all components
>>> have the green symbol (*hop.error.visual.is.ok.but.logging.is.error.jpg*
>>> )
>>>
>>> The field that caused the error, "cpf", in the source table is a
>>> numeric(15) and in the destination table it is of type "text", that is, at
>>> some point the type conversion is necessary.
>>>
>>> In Pentaho Data Integration (aka kettle) I didn't get this type of
>>> error. The only difference is that in it, the "table input" component has
>>> the "Enable Lazy Conversion" option, which I enabled. In Apache Hop,
>>> however, this option doesn't exist, but it is listed in the official
>>> documentation (*hop.table.input.where.is.enable.lazy.conversion.jpg*)
>>>
>>> *Questions*:
>>> - Does the "Enable Lazy Conversion" option exist or not in "Table
>>> Input"? If it doesn't exist, how would I perform the type conversion. And
>>> if it does, why doesn't it appear in my window? Could it be a bug?
>>> - Is it normal for this error to be thrown in the logging and not
>>> reflected in the visual part of the components?
>>>
>>>
>>>
>>>

Reply via email to