跳转至

🐦‍⬛ ailab100.pdbc.datawarehouse

alcedo_pdbc.datawarehouse

@File :init.py @Time :2025-7-16 14:38 @Author:AI Lab Morgan

Classes

alcedo_pdbc.datawarehouse.BigQuery
Functions
alcedo_pdbc.datawarehouse.BigQuery.download_as_file(query, filename) method descriptor

BigQuery.download_as_file(self, str query: str, str filename: str) -> None

Takes query, filename as arguments and download the data as file

Parameters:

Name Type Description Default
query str

select query

required
filename str

filename to save the file

required

Returns:

Name Type Description
file ``CSV``、``Excel``、 ``JSON``、``HTML``、 ``HDF5``、 ``Feather``、 ``Parquet`` 、``Apache Avro``

根据参数filename文件类型导出文件保存至指定的目录下

alcedo_pdbc.datawarehouse.BigQuery.read_as_dataframe(query, return_type='pandas') method descriptor

BigQuery.read_as_dataframe(self, str query: str, str return_type: str = 'pandas')

Takes query as the arguments and return the dataframe

Parameters:

Name Type Description Default
query str

select query

required
return_type str

which dataframe you want to return (pandas, polars, dask etc). Defaults to

'pandas'

Returns:

Name Type Description
DataFrame ``Pandas``、``Polars`` or ``Dask``

根据 return_type 参数返回对应的数据帧 Dataframe

alcedo_pdbc.datawarehouse.BigQuery.write_dataframe(df, table_name, project_id, if_exists='append') method descriptor

BigQuery.write_dataframe(self, df, str table_name: str, str project_id: str, str if_exists: str = 'append') -> None

Takes dataframe, table name, project id as arguments and write the dataframe to BigQuery

Parameters:

Name Type Description Default
df DataFrame

Dataframe which need to be loaded

required
table_name str

table name

required
project_id str

project id

required
if_exists str

operation to do if the table exists. Defaults to 'append'.

'append'
alcedo_pdbc.datawarehouse.Redshift

Bases: alcedo_pdbc.sql.sql_utils.DBConnector

Functions
alcedo_pdbc.datawarehouse.Redshift.write_dataframe(df, table_name, database=None, if_exists='append', index=False) method descriptor

Redshift.write_dataframe(self, df, str table_name: str, str database: str = None, str if_exists: str = 'append', index=False)

Takes dataframe, table name as arguments and write the dataframe to Redshift

Parameters:

Name Type Description Default
df DataFrame

Dataframe which need to be loaded

required
table_name str

table name

required
database str

database name. Defaults to None.

None
if_exists str

operation to do if the table exists. Defaults to 'append'.

'append'
index bool

Write DataFrame index as a column. Defaults to False.

False
alcedo_pdbc.datawarehouse.SnowFlake
Functions
alcedo_pdbc.datawarehouse.SnowFlake.download_as_file(query, filename, database=None, schema=None, protocol='https') method descriptor

SnowFlake.download_as_file(self, str query: str, str filename: str, str database: str = None, str schema: str = None, str protocol: str = 'https') -> None

Takes query, filename as arguments and download the data as file

Parameters:

Name Type Description Default
query str

select query

required
filename str

filename to save the file

required
database str

database name, if None, it take it from config. Defaults to None.

None
schema str

schema name, if None, it take it from config. Defaults to None.

None

Returns:

Name Type Description
file ``CSV``、``Excel``、 ``JSON``、``HTML``、 ``HDF5``、 ``Feather``、 ``Parquet`` 、``Apache Avro``

根据参数filename文件类型导出文件保存至指定的目录下

alcedo_pdbc.datawarehouse.SnowFlake.read_as_dataframe(query, database=None, schema=None, protocol='https', return_type='pandas') method descriptor

SnowFlake.read_as_dataframe(self, str query: str, str database: str = None, str schema: str = None, str protocol: str = 'https', str return_type: str = 'pandas')

Takes query as arguments and return dataframe

Parameters:

Name Type Description Default
query str

select query

required
database str

database name, if None, it take it from config. Defaults to None.

None
schema str

schema name, if None, it take it from config. Defaults to None.

None
protocol str

protocol Defaults to 'https'.

'https'
return_type str

which dataframe you want to return (pandas, polars, dask etc). Defaults to

'pandas'

Returns:

Name Type Description
DataFrame ``Pandas``、``Polars`` or ``Dask``

根据 return_type 参数返回对应的数据帧 Dataframe

alcedo_pdbc.datawarehouse.SnowFlake.write_dataframe(df, table_name, database=None, schema=None, protocol='https') method descriptor

SnowFlake.write_dataframe(self, df, str table_name: str, str database: str = None, str schema: str = None, str protocol: str = 'https')

Takes dataframe, table name as arguments and write the dataframe to SnowFlake

Parameters:

Name Type Description Default
df Dataframe

Dataframe which need to be loaded

required
table_name str

table name

required
database str

database name. Defaults to None.

None
schema str

schema name. Defaults to None.

None
protocol str

protocol used. Defaults to 'https'.

'https'
alcedo_pdbc.datawarehouse.StarRocks

StarRocks class create the ligo starrocks object, through which you can able to read, write, download data from StarRocks.

Parameters:

Name Type Description Default
config dict

Automatically loaded from the config file (yaml)

required
Functions
alcedo_pdbc.datawarehouse.StarRocks.download_as_file(query, filename, database=None) method descriptor

StarRocks.download_as_file(self, str query: str, str filename: str, str database: str = None) -> None

Takes query, filename as arguments and download the data as file

Parameters:

Name Type Description Default
query str

select query

required
filename str

filename to save the file

required
database str

database name, if None, it take it from config. Defaults to None.

None

Returns:

Name Type Description
file ``CSV``、``Excel``、 ``JSON``、``HTML``、 ``HDF5``、 ``Feather``、 ``Parquet`` 、``Apache Avro``

根据参数filename文件类型导出文件保存至指定的目录下

alcedo_pdbc.datawarehouse.StarRocks.read_as_dataframe(query, database=None, return_type='pandas') method descriptor

StarRocks.read_as_dataframe(self, str query: str, str database: str = None, str return_type: str = 'pandas')

Takes query as argument and return a dataframe

Parameters:

Name Type Description Default
query str

select query

required
database str

database name, if None, it take it from config. Defaults to None.

None
return_type str

which dataframe you want to return (pandas, polars, dask etc). Defaults to 'pandas'. Defaults to 'pandas'.

'pandas'

Returns:

Name Type Description
DataFrame ``Pandas``、``Polars`` or ``Dask``

根据 return_type 参数返回对应的数据帧 Dataframe

alcedo_pdbc.datawarehouse.StarRocks.write_dataframe(df, table_name, database=None, if_exists='append', index=False) method descriptor

StarRocks.write_dataframe(self, df, str table_name: str, str database: str = None, str if_exists: str = 'append', index=False)

Takes dataframe, table name as arguments and write the dataframe to StarRocks

Parameters:

Name Type Description Default
df DataFrame

Dataframe which need to be loaded

required
table_name str

table name

required
database str

database name. Defaults to None.

None
if_exists str

operation to do if the table exists. Defaults to 'append'.

'append'
index bool

Write DataFrame index as a column. Defaults to False.

False