Hi Rajnil,

Sorry for the multiple emails. It seems you are getting the
ModuleNotFoundError error was curious, have you tried using the
below-mentioned solution mentioned in the readme file?

Below is the link:-
https://github.com/GoogleCloudDataproc/spark-bigquery-connector#bignumeric-support

Also please find the code block solution.

if the code throws ModuleNotFoundError, please add the following code
before reading the BigNumeric data.

try:
    import pkg_resources

    pkg_resources.declare_namespace(__name__)
except ImportError:
    import pkgutil

    __path__ = pkgutil.extend_path(__path__, __name__)

Thanks,

Atheeth


On Fri, 3 Mar 2023 at 16:25, Atheeth SH <atheeth.e...@gmail.com> wrote:

> Hi Rajnil,
>
> Just curious, what version of spark-bigquery-connector are you using?
>
> Thanks,
> Atheeth
>
> On Sat, 25 Feb 2023 at 23:48, Mich Talebzadeh <mich.talebza...@gmail.com>
> wrote:
>
>> sounds like it is cosmetric. The important point is that if the data
>> stored in GBQ is valid?
>>
>>
>> THT
>>
>>
>>    view my Linkedin profile
>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>
>>
>>  https://en.everybodywiki.com/Mich_Talebzadeh
>>
>>
>>
>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>> any loss, damage or destruction of data or any other property which may
>> arise from relying on this email's technical content is explicitly
>> disclaimed. The author will in no case be liable for any monetary damages
>> arising from such loss, damage or destruction.
>>
>>
>>
>>
>> On Sat, 25 Feb 2023 at 18:12, Rajnil Guha <rajnil94.g...@gmail.com>
>> wrote:
>>
>>> Hi All,
>>>
>>> I had created an issue on Stackoverflow(linked below) a few months back
>>> about issues while handling bignumeric type values of BigQuery in Spark.
>>>
>>> link
>>> <https://stackoverflow.com/questions/74719503/getting-error-while-reading-bignumeric-data-type-from-a-bigquery-table-using-apa>
>>>
>>> On Fri, Feb 24, 2023 at 3:54 PM Mich Talebzadeh <
>>> mich.talebza...@gmail.com> wrote:
>>>
>>>> Hi Nidhi,
>>>>
>>>> can you create a BigQuery table with a  bignumeric and numeric column
>>>> types, add a few lines and try to read into spark. through DF
>>>>
>>>> and do
>>>>
>>>>
>>>> df.printSchema()
>>>>
>>>> df.show(5,False)
>>>>
>>>>
>>>> HTH
>>>>
>>>>
>>>>    view my Linkedin profile
>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>
>>>>
>>>>  https://en.everybodywiki.com/Mich_Talebzadeh
>>>>
>>>>
>>>>
>>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>>> any loss, damage or destruction of data or any other property which may
>>>> arise from relying on this email's technical content is explicitly
>>>> disclaimed. The author will in no case be liable for any monetary damages
>>>> arising from such loss, damage or destruction.
>>>>
>>>>
>>>>
>>>>
>>>> On Fri, 24 Feb 2023 at 02:47, nidhi kher <kherni...@gmail.com> wrote:
>>>>
>>>>> Hello,
>>>>>
>>>>>
>>>>> I am facing below issue in pyspark code:
>>>>>
>>>>> We are running spark code using dataproc serverless batch in google
>>>>> cloud platform. Spark code is causing issue while writing the data to
>>>>> bigquery table. In bigquery table , few of the columns have datatype as
>>>>> bignumeric and spark code is changing the datatype from bignumeric to
>>>>> numeric while writing the data. We need datatype to be kept as bignumeric
>>>>> only as we need data of 38,20 precision.
>>>>>
>>>>>
>>>>> Can we cast a column to bignumeric in spark sql dataframe like below
>>>>> code for decimal:
>>>>>
>>>>>
>>>>> df= spark.sql("""SELECT cast(col1 as decimal(38,20)) as col1 from
>>>>> table1""")
>>>>>
>>>>> Spark version :3.3
>>>>>
>>>>> Pyspark version : 1.1
>>>>>
>>>>>
>>>>> Regards,
>>>>>
>>>>> Nidhi
>>>>>
>>>>

Reply via email to