
    4i_                       S r SSKJr  SSKrSSKrSSKrSSKJrJrJ	r	  SSK
JrJr  SSKJr  SSKJr  SSKJrJr  SS	KJr  SS
KJr  SSKJrJr  SSKJr  SSKJrJrJ r J!r!J"r"  \(       a  SSK#J$r$J%r%J&r&J'r'J(r(J)r)  SS jr*   S           SS jjr+ " S S5      r, " S S\,5      r- " S S\,5      r.       S                 SS jjr/\" S5      SSS\R`                  SSS4                 S S jj5       r1g)!zparquet compat    )annotationsN)TYPE_CHECKINGAnyLiteral)catch_warningsfilterwarnings)lib)import_optional_dependency)AbstractMethodErrorPandas4Warning)
set_module)check_dtype_backend)	DataFrame
get_option)arrow_table_to_pandas)	IOHandles
get_handleis_fsspec_urlis_urlstringify_path)DtypeBackendFilePathParquetCompressionOptions
ReadBufferStorageOptionsWriteBufferBaseImplc                2   U S:X  a  [        S5      n U S:X  a-  [        [        /nSnU H  n U" 5       s  $    [        SU 35      eU S:X  a
  [        5       $ U S:X  a
  [        5       $ [        S	5      e! [         a  nUS[	        U5      -   -  n SnAMi  SnAff = f)
zreturn our implementationautozio.parquet.engine z
 - NzUnable to find a usable engine; tried using: 'pyarrow', 'fastparquet'.
A suitable version of pyarrow or fastparquet is required for parquet support.
Trying to import the above resulted in these errors:pyarrowfastparquetz.engine must be one of 'pyarrow', 'fastparquet')r   PyArrowImplFastParquetImplImportErrorstr
ValueError)engineengine_classes
error_msgsengine_classerrs        S/var/www/html/dynamic-report/venv/lib/python3.13/site-packages/pandas/io/parquet.py
get_enginer.   4   s    /0%7
*L1#~% + C l
 	
 }	=	   
E
FF%  1gC00
1s   A00
B:BBc                   [        U 5      nUb  [        SSS9n[        SSS9nUb-  [        XR                  5      (       a  U(       a  [	        S5      eOIUb%  [        XR
                  R                  5      (       a  O![        S[        U5      R                   35      e[        U5      (       aq  Ucn  Uc4  [        S5      n[        S5      n UR                  R                  U 5      u  pUc3  [        S5      nUR                  R                  " U40 U=(       d    0 D6u  pO(U(       a!  [!        U5      (       a  US	:w  a  [        S
5      eSn	U(       dY  U(       dR  [        U["        5      (       a=  [$        R&                  R)                  U5      (       d  [+        XSSUS9n	SnU	R,                  nXYU4$ ! [        UR                  4 a     Nf = f)zFile handling for PyArrow.Nz
pyarrow.fsignore)errorsfsspecz8storage_options not supported with a pyarrow FileSystem.z9filesystem must be a pyarrow or fsspec FileSystem, not a r!   rbz8storage_options passed with buffer, or non-supported URLFis_textstorage_options)r   r
   
isinstance
FileSystemNotImplementedErrorspecAbstractFileSystemr'   type__name__r   from_uri	TypeErrorArrowInvalidcore	url_to_fsr   r&   ospathisdirr   handle)
rD   fsr6   modeis_dirpath_or_handlepa_fsr2   pahandless
             r-   _get_path_or_handlerN   V   s    $D)N	~*<I+HXFB0@0@!A!A)N   Jr;;3Q3Q$R$Rb**+-  ^$$"+I6B.|<E%*%5%5%>%>t%D" :/9F!'!6!6"#2#8b"B 
&"8"8DDL STTG~s++n--
 %
  B&&7 r/ s   	F+ +GGc                  @    \ rS rSr\SS j5       rSS jrSS	S jjrSrg)
r      c                D    [        U [        5      (       d  [        S5      eg )Nz+to_parquet only supports IO with DataFrames)r7   r   r'   )dfs    r-   validate_dataframeBaseImpl.validate_dataframe   s     "i((JKK )    c                    [        U 5      eNr   )selfrR   rD   compressionkwargss        r-   writeBaseImpl.write       !$''rU   Nc                    [        U 5      erW   rX   )rY   rD   columnsr[   s       r-   readBaseImpl.read   r^   rU    )rR   r   returnNonerW   )rd   r   )	r=   
__module____qualname____firstlineno__staticmethodrS   r\   ra   __static_attributes__rc   rU   r-   r   r      s%    L L(( (rU   c                      \ rS rSrSS jr     S             S	S jjrSS\R                  SSS4       S
S jjrSr	g)r#      c                4    [        SSS9  SS KnSS KnXl        g )Nr!   z(pyarrow is required for parquet support.extrar   )r
   pyarrow.parquet(pandas.core.arrays.arrow.extension_typesapi)rY   r!   pandass      r-   __init__PyArrowImpl.__init__   s    "G	
 	 	8rU   Nc                   U R                  U5        SUR                  SS 5      0n	Ub  XIS'   U R                  R                  R                  " U40 U	D6n
UR
                  (       aO  S[        R                  " UR
                  5      0nU
R                  R                  n0 UEUEnU
R                  U5      n
[        UUUSUS LS9u  pn[        U[        R                  5      (       a|  [        US5      (       ak  [        UR                   ["        [$        45      (       aF  [        UR                   [$        5      (       a  UR                   R'                  5       nOUR                   n Ub-  U R                  R(                  R*                  " U
U4UUUS.UD6  O+U R                  R(                  R,                  " U
U4UUS.UD6  Ub  UR/                  5         g g ! Ub  UR/                  5         f f = f)	Nschemapreserve_indexPANDAS_ATTRSwb)r6   rH   rI   name)rZ   partition_cols
filesystem)rZ   r}   )rS   poprr   Tablefrom_pandasattrsjsondumpsrw   metadatareplace_schema_metadatarN   r7   ioBufferedWriterhasattrr{   r&   bytesdecodeparquetwrite_to_datasetwrite_tableclose)rY   rR   rD   rZ   indexr6   r|   r}   r[   from_pandas_kwargstabledf_metadataexisting_metadatamerged_metadatarJ   rM   s                   r-   r\   PyArrowImpl.write   s    	#.6

8T8R-S38/0**2D1CD88)4::bhh+?@K % 5 5B!2BkBO11/BE.A+!-/
+ ~r'8'899//>..e==.--u55!/!4!4!;!;!=!/!4!4	 )  11" !,#1)    ,," !,)	
  " #w" #s   "AG G(c                2   SUS'   [        UUUSS9u  pn U R                  R                  R                  " U	4UUUS.UD6n[	        5          [        SS[        5        [        UUUS9nS S S 5        UR                  R                  (       aN  S	UR                  R                  ;   a4  UR                  R                  S	   n[        R                  " U5      Wl        WU
b  U
R                  5         $ $ ! , (       d  f       N= f! U
b  U
R                  5         f f = f)
NTuse_pandas_metadatar3   )r6   rH   )r`   r}   filtersr0   make_block is deprecated)dtype_backendto_pandas_kwargss   PANDAS_ATTRS)rN   rr   r   
read_tabler   r   r   r   rw   r   r   loadsr   r   )rY   rD   r`   r   r   r6   r}   r   r[   rJ   rM   pa_tableresultr   s                 r-   ra   PyArrowImpl.read   s    )-$%.A+	/
+	 xx''22%	
 H  !."
 /"/%5 " ''"hoo&>&>>"*//":":?"KK#'::k#:FL" #% "!$ " #s$   5D  C/(A2D  /
C=9D   Drr   rd   re   snappyNNNN)rR   r   rD   zFilePath | WriteBuffer[bytes]rZ   r   r   bool | Noner6   StorageOptions | Noner|   list[str] | Nonerd   re   )r   DtypeBackend | lib.NoDefaultr6   r   r   zdict[str, Any] | Nonerd   r   )
r=   rf   rg   rh   rt   r\   r	   
no_defaultra   rj   rc   rU   r-   r#   r#      s    	 2:!15+/@ @  ,@  /	@ 
 @  /@  )@  
@ J 69nn1526. 
 4.  /.  0.  
.  . rU   r#   c                  f    \ rS rSrSS jr     S       S	S jjr     S
     SS jjrSrg)r$   i!  c                $    [        SSS9nXl        g )Nr"   z,fastparquet is required for parquet support.rn   )r
   rr   )rY   r"   s     r-   rt   FastParquetImpl.__init__"  s     1!O
 rU   Nc                  ^^	 U R                  U5        SU;   a  Ub  [        S5      eSU;   a  UR                  S5      nUb  SUS'   Ub  [        S5      e[	        U5      n[        U5      (       a  [        S5      m	U	U4S jUS'   OT(       a  [        S	5      e[        S
S9   U R                  R                  " UU4UUUS.UD6  S S S 5        g ! , (       d  f       g = f)Npartition_onzYCannot use both partition_on and partition_cols. Use partition_cols for partitioning datahivefile_scheme9filesystem is not implemented for the fastparquet engine.r2   c                Z   > TR                   " U S40 T=(       d    0 D6R                  5       $ )Nrz   )open)rD   _r2   r6   s     r-   <lambda>'FastParquetImpl.write.<locals>.<lambda>M  s,    &++d3.4"3df3rU   	open_withz?storage_options passed with file object or non-fsspec file pathT)record)rZ   write_indexr   )
rS   r'   r~   r9   r   r   r
   r   rr   r\   )
rY   rR   rD   rZ   r   r|   r6   r}   r[   r2   s
         `  @r-   r\   FastParquetImpl.write*  s     	#V#(BK  V##ZZ7N%$*F=!!%K 
 d#/9F#F; Q  4(HHNN (!+  )((s   !#C
Cc                b   0 nUR                  S[        R                  5      n	SUS'   U	[        R                  La  [        S5      eUb  [	        S5      eUb  [	        S5      e[        U5      nS n
[        U5      (       a6  [        S5      nUR                  " US40 U=(       d    0 D6R                  US	'   OQ[        U[        5      (       a<  [        R                  R                  U5      (       d  [        USSUS
9n
U
R                   n U R"                  R$                  " U40 UD6n['        5          [)        SS[*        5        UR,                  " SX#S.UD6sS S S 5        U
b  U
R/                  5         $ $ ! , (       d  f       O= f U
b  U
R/                  5         g g ! U
b  U
R/                  5         f f = f)Nr   Fpandas_nullszHThe 'dtype_backend' argument is not supported for the fastparquet enginer   z?to_pandas_kwargs is not implemented for the fastparquet engine.r2   r3   rG   r4   r0   r   )r`   r   rc   )r~   r	   r   r'   r9   r   r   r
   r   rG   r7   r&   rC   rD   rE   r   rF   rr   ParquetFiler   r   r   	to_pandasr   )rY   rD   r`   r   r6   r}   r   r[   parquet_kwargsr   rM   r2   parquet_files                r-   ra   FastParquetImpl.read_  s    *,

?CNNC).~&.%  !%K  '%Q  d#/9F#);;tT#Uo>SQS#U#X#XN4 c""277==+>+> !dE?G >>D	 88//GGL!."
 $-- #8> "! " # "!! " #w" #s$   'F .%E1	F 1
E?;F F.r   r   r   )rR   r   rZ   z*Literal['snappy', 'gzip', 'brotli'] | Noner6   r   rd   re   )NNNNN)r6   r   r   dict | Nonerd   r   )r=   rf   rg   rh   rt   r\   ra   rj   rc   rU   r-   r$   r$   !  s     CK1533 @	3 /3 
3p 15(,7 
 /7  &7  
7  7 rU   r$   r   c           	        [        U[        5      (       a  U/n[        U5      n	Uc  [        R                  " 5       OUn
U	R
                  " U U
4UUUUUS.UD6  Uc1  [        U
[        R                  5      (       d   eU
R                  5       $ g)a   
Write a DataFrame to the parquet format.

Parameters
----------
df : DataFrame
path : str, path object, file-like object, or None, default None
    String, path object (implementing ``os.PathLike[str]``), or file-like
    object implementing a binary ``write()`` function. If None, the result
    is returned as bytes. If a string, it will be used as Root Directory
    path when writing a partitioned dataset. The engine fastparquet does
    not accept file-like objects.
engine : {{'auto', 'pyarrow', 'fastparquet'}}, default 'auto'
    Parquet library to use. If 'auto', then the option
    ``io.parquet.engine`` is used. The default ``io.parquet.engine``
    behavior is to try 'pyarrow', falling back to 'fastparquet' if
    'pyarrow' is unavailable.

    When using the ``'pyarrow'`` engine and no storage options are provided
    and a filesystem is implemented by both ``pyarrow.fs`` and ``fsspec``
    (e.g. "s3://"), then the ``pyarrow.fs`` filesystem is attempted first.
    Use the filesystem keyword with an instantiated fsspec filesystem
    if you wish to use its implementation.
compression : {{'snappy', 'gzip', 'brotli', 'lz4', 'zstd', None}},
    default 'snappy'. Name of the compression to use. Use ``None``
    for no compression.
index : bool, default None
    If ``True``, include the dataframe's index(es) in the file output. If
    ``False``, they will not be written to the file.
    If ``None``, similar to ``True`` the dataframe's index(es)
    will be saved. However, instead of being saved as values,
    the RangeIndex will be stored as a range in the metadata so it
    doesn't require much space and is faster. Other indexes will
    be included as columns in the file output.
partition_cols : str or list, optional, default None
    Column names by which to partition the dataset.
    Columns are partitioned in the order they are given.
    Must be None if path is not a string.
storage_options : dict, optional
    Extra options that make sense for a particular storage connection, e.g.
    host, port, username, password, etc. For HTTP(S) URLs the key-value
    pairs are forwarded to ``urllib.request.Request`` as header options.
    For other URLs (e.g. starting with "s3://", and "gcs://") the
    key-value pairs are forwarded to ``fsspec.open``. Please see ``fsspec``
    and ``urllib`` for more details, and for more examples on storage
    options refer `here <https://pandas.pydata.org/docs/user_guide/io.html?
    highlight=storage_options#reading-writing-remote-files>`_.
filesystem : fsspec or pyarrow filesystem, default None
    Filesystem object to use when reading the parquet file. Only implemented
    for ``engine="pyarrow"``.

    .. versionadded:: 2.1.0

**kwargs
    Additional keyword arguments passed to the engine:

    * For ``engine="pyarrow"``: passed to :func:`pyarrow.parquet.write_table`
      or :func:`pyarrow.parquet.write_to_dataset` (when using partition_cols)
    * For ``engine="fastparquet"``: passed to :func:`fastparquet.write`

Returns
-------
bytes if no path argument is provided else None
N)rZ   r   r|   r6   r}   )r7   r&   r.   r   BytesIOr\   getvalue)rR   rD   r(   rZ   r   r6   r|   r}   r[   implpath_or_bufs              r-   
to_parquetr     s    V .#&&()fDAESWKJJ
	  %'	 	 |+rzz2222##%%rU   rs   c           
     b    [        U5      n	[        U5        U	R                  " U 4UUUUUUS.UD6$ )aQ  
Load a parquet object from the file path, returning a DataFrame.

The function automatically handles reading the data from a parquet file
and creates a DataFrame with the appropriate structure.

Parameters
----------
path : str, path object or file-like object
    String, path object (implementing ``os.PathLike[str]``), or file-like
    object implementing a binary ``read()`` function.
    The string could be a URL. Valid URL schemes include http, ftp, s3,
    gs, and file. For file URLs, a host is expected. A local file could be:
    ``file://localhost/path/to/table.parquet``.
    A file URL can also be a path to a directory that contains multiple
    partitioned parquet files. Both pyarrow and fastparquet support
    paths to directories as well as file URLs. A directory path could be:
    ``file://localhost/path/to/tables`` or ``s3://bucket/partition_dir``.
engine : {{'auto', 'pyarrow', 'fastparquet'}}, default 'auto'
    Parquet library to use. If 'auto', then the option
    ``io.parquet.engine`` is used. The default ``io.parquet.engine``
    behavior is to try 'pyarrow', falling back to 'fastparquet' if
    'pyarrow' is unavailable.

    When using the ``'pyarrow'`` engine and no storage options are provided
    and a filesystem is implemented by both ``pyarrow.fs`` and ``fsspec``
    (e.g. "s3://"), then the ``pyarrow.fs`` filesystem is attempted first.
    Use the filesystem keyword with an instantiated fsspec filesystem
    if you wish to use its implementation.
columns : list, default=None
    If not None, only these columns will be read from the file.
storage_options : dict, optional
    Extra options that make sense for a particular storage connection, e.g.
    host, port, username, password, etc. For HTTP(S) URLs the key-value
    pairs are forwarded to ``urllib.request.Request`` as header options.
    For other URLs (e.g. starting with "s3://", and "gcs://") the
    key-value pairs are forwarded to ``fsspec.open``. Please see ``fsspec``
    and ``urllib`` for more details, and for more examples on storage
    options refer `here <https://pandas.pydata.org/docs/user_guide/io.html?
    highlight=storage_options#reading-writing-remote-files>`_.
dtype_backend : {{'numpy_nullable', 'pyarrow'}}
    Back-end data type applied to the resultant :class:`DataFrame`
    (still experimental). If not specified, the default behavior
    is to not use nullable data types. If specified, the behavior
    is as follows:

    * ``"numpy_nullable"``: returns nullable-dtype-backed :class:`DataFrame`
    * ``"pyarrow"``: returns pyarrow-backed nullable
      :class:`ArrowDtype` :class:`DataFrame`

    .. versionadded:: 2.0

filesystem : fsspec or pyarrow filesystem, default None
    Filesystem object to use when reading the parquet file. Only implemented
    for ``engine="pyarrow"``.

    .. versionadded:: 2.1.0

filters : List[Tuple] or List[List[Tuple]], default None
    To filter out data.
    Filter syntax: [[(column, op, val), ...],...]
    where op is [==, =, >, >=, <, <=, !=, in, not in]
    The innermost tuples are transposed into a set of filters applied
    through an `AND` operation.
    The outer list combines these sets of filters through an `OR`
    operation.
    A single list of tuples can also be used, meaning that no `OR`
    operation between set of filters is to be conducted.

    Using this argument will NOT result in row-wise filtering of the final
    partitions unless ``engine="pyarrow"`` is also specified.  For
    other engines, filtering is only performed at the partition level, that is,
    to prevent the loading of some row-groups and/or files.

    .. versionadded:: 2.1.0

to_pandas_kwargs : dict | None, default None
    Keyword arguments to pass through to :func:`pyarrow.Table.to_pandas`
    when ``engine="pyarrow"``.

    .. versionadded:: 3.0.0

**kwargs
    Additional keyword arguments passed to the engine:

    * For ``engine="pyarrow"``: passed to :func:`pyarrow.parquet.read_table`
    * For ``engine="fastparquet"``: passed to
      :meth:`fastparquet.ParquetFile.to_pandas`

Returns
-------
DataFrame
    DataFrame based on parquet file.

See Also
--------
DataFrame.to_parquet : Create a parquet object that serializes a DataFrame.

Examples
--------
>>> original_df = pd.DataFrame({"foo": range(5), "bar": range(5, 10)})
>>> original_df
   foo  bar
0    0    5
1    1    6
2    2    7
3    3    8
4    4    9
>>> df_parquet_bytes = original_df.to_parquet()
>>> from io import BytesIO
>>> restored_df = pd.read_parquet(BytesIO(df_parquet_bytes))
>>> restored_df
   foo  bar
0    0    5
1    1    6
2    2    7
3    3    8
4    4    9
>>> restored_df.equals(original_df)
True
>>> restored_bar = pd.read_parquet(BytesIO(df_parquet_bytes), columns=["bar"])
>>> restored_bar
    bar
0    5
1    6
2    7
3    8
4    9
>>> restored_bar.equals(original_df[["bar"]])
True

The function uses `kwargs` that are passed directly to the engine.
In the following example, we use the `filters` argument of the pyarrow
engine to filter the rows of the DataFrame.

Since `pyarrow` is the default engine, we can omit the `engine` argument.
Note that the `filters` argument is implemented by the `pyarrow` engine,
which can benefit from multithreading and also potentially be more
economical in terms of memory.

>>> sel = [("foo", ">", 2)]
>>> restored_part = pd.read_parquet(BytesIO(df_parquet_bytes), filters=sel)
>>> restored_part
    foo  bar
0    3    8
1    4    9
)r`   r   r6   r   r}   r   )r.   r   ra   )
rD   r(   r`   r6   r   r}   r   r   r[   r   s
             r-   read_parquetr     sL    @ fD&99	'#)	 	 	rU   )r(   r&   rd   r   )Nr3   F)rD   z1FilePath | ReadBuffer[bytes] | WriteBuffer[bytes]rG   r   r6   r   rH   r&   rI   boolrd   zVtuple[FilePath | ReadBuffer[bytes] | WriteBuffer[bytes], IOHandles[bytes] | None, Any])Nr   r   NNNN)rR   r   rD   z$FilePath | WriteBuffer[bytes] | Noner(   r&   rZ   r   r   r   r6   r   r|   r   r}   r   rd   zbytes | None)rD   zFilePath | ReadBuffer[bytes]r(   r&   r`   r   r6   r   r   r   r}   r   r   z&list[tuple] | list[list[tuple]] | Noner   r   rd   r   )2__doc__
__future__r   r   r   rC   typingr   r   r   warningsr   r   pandas._libsr	   pandas.compat._optionalr
   pandas.errorsr   r   pandas.util._decoratorsr   pandas.util._validatorsr   rs   r   r   pandas.io._utilr   pandas.io.commonr   r   r   r   r   pandas._typingr   r   r   r   r   r   r.   rN   r   r#   r$   r   r   r   rc   rU   r-   <module>r      s    " 	  	 

  > / 7
 2   GJ .2<'
;<'<' +<' 	<'
 <'<'~
( 
(| ( | ~u h u t 26-5-1'+``
.` ` +	`
 ` +` %` ` `F H  $-125..6:$(k
&kk k +	k
 0k k 4k "k k krU   