Skip to main content

A sample test package

Project description

afl-ai-utils

rm -rf build dist 
python3 setup.py sdist bdist_wheel
twine upload --repository pypi dist/* 

Installation

    pip install afl-ai-utils

Usage

Slack Alerting

from afl_ai_utils.slack_alerts import send_slack_alert 
send_slack_alert(info_alert_slack_webhook_url=None, red_alert_slack_webhook_url=None, slack_red_alert_userids=None, payload=None, is_red_alert=False)

    """Send a Slack message to a channel via a webhook.

Args:
    info_alert_slack_webhook_url(str): Infor slack channel url
    red_alert_slack_webhook_url(str): red alert channel url
    slack_red_alert_userids (list): userid's to mention in slack for red alert notification
    payload (dict): Dictionary containing Slack message, i.e. {"text": "This is a test"}
    is_red_alert (bool): Full Slack webhook URL for your chosen channel.

Returns:
    HTTP response code, i.e. <Response [503]>
"""

BigQuery Dataframe to BigQuery and get result in Datafeame

    def dump_dataframe_to_bq_table(self, dataframe: pd.DataFrame, schema_cols_type: dict, table_id: str, mode: str):

    >>> from afl_ai_utils.bigquery_utils import BigQuery
    >>> bq = BigQuery("keys.json")
    >>> bq.write_insights_to_bq_table(dataframe=None, schema=None, table_id=None, mode=None)
    
    
    """Insert a dataframe to bigquery

    Args:
        dataframe(pandas dataframe): for dataframe to be dumped to bigquery
        schema(BigQuery.Schema ): ex:
        schema_cols_type: {"date_start":"date", "id": "integer", "name": "string"}
        table_id (list): table_id in which dataframe need to be inserted e.g project_id.dataset.table_name = table_id
        mode(str): To append or replace the table - e.g mode = "append"  or mode="replace"
    Returns:
        returns as success message with number of inserted rows and table name
    """

Execute any query to BigQuery

    def execute_query(self, query):

    >>> from afl_ai_utils.bigquery_utils import BigQuery
    >>> bq = BigQuery("keys.json")
    >>> df = bq.execute_query(query = "SELECT * FROM TABLE")
    
    
    """
    Args:
        query (query of any type SELECT/INSERT/DELETE ) 
    Returns:
        returns dataframe of execute query result
    """

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

afl-ai-utils-0.6.8.tar.gz (20.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

afl_ai_utils-0.6.8-py3-none-any.whl (24.4 kB view details)

Uploaded Python 3

File details

Details for the file afl-ai-utils-0.6.8.tar.gz.

File metadata

  • Download URL: afl-ai-utils-0.6.8.tar.gz
  • Upload date:
  • Size: 20.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.6

File hashes

Hashes for afl-ai-utils-0.6.8.tar.gz
Algorithm Hash digest
SHA256 522681c6c13818ae32bc933fc14ebc97a93c23bdeae09a64b6f2cab876e99378
MD5 8f6bef3a268f4f204a0b3a1f22de2ef6
BLAKE2b-256 d83b3abb9f4a88f8779a28fc348407109767d2b28c8c7510ade1187caaaf62e1

See more details on using hashes here.

File details

Details for the file afl_ai_utils-0.6.8-py3-none-any.whl.

File metadata

  • Download URL: afl_ai_utils-0.6.8-py3-none-any.whl
  • Upload date:
  • Size: 24.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.6

File hashes

Hashes for afl_ai_utils-0.6.8-py3-none-any.whl
Algorithm Hash digest
SHA256 de777673c3f92ba5bb464a2ed82d10cf2b76b3cffef4631a1f507ef355a58342
MD5 e04ee9bd3ad11559472905587e7bc5cc
BLAKE2b-256 5c9515f0579257382b6fb007f56dd77d94fe06b0ffc801e827f2579409043b2f

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page