Skip to main content

A productive library to extract data from Azure Devops and apply agile metrics.

Project description

Azure Devops PySpark: A productive library to extract data from Azure Devops and apply agile metrics.

What is it?

Azure Devops PySpark is a Python package that provides the most productive way to extract data from Azure Devops and build agile metrics. It runs on PySpark, enabling all the features the technology makes available.

Main Features

  • Get authenticated quickly and simply.

  • All columns of the project are automatically mapped, use the filter_columns method to filter only those you want to form your dataframes with.

  • SparkSession already created in spark variable.

  • Get all your organization's backlogs with the method backlog.

  • Get all your organization's items with the method items.

  • Get all your organization's iterations with the method iterations.

  • Get all your organization's members with the method members.

  • Get all your organization's tags with the method tags.

  • Get all your organization's teams with the method teams.

  • Explore the simplicity of Agile class to build powerful metrics for your organization.

Quick example

from AzureDevopsPySpark import Azure, Agile
from pyspark.sql.functions import datediff #use in agile metrics

devops = Azure('ORGANIZATION', 'PROJECT', 'PERSONAL ACCESS TOKENS')
## Filter columns
devops.filter_columns([
    'IterationPath', 'Id', 'State', 'WorkItemType',
    'CreatedDate', 'ClosedDate', 'Iteration_Start_Date', 'Iteration_End_Date'
])


## Pyspark Dataframe
df_backlog = devops.backlog()
df_items = devops.items()
df_iterations = devops.iterations()
df_members = devops.members()
df_tags = devops.tags()
df_teams = devops.teams()
## Agile Metrics
agile = Agile()

## A new dataframe
df_agil = df.items.join(df_iterations, 'IterationPath')

## Metrics

## Average time between CreatedDate and ClosedDate of items in the last 90 days.
lead_time = agile.avg(
    df=df_agil,
    ref=[datediff, 'ClosedDate', 'CreatedDate'], # The day difference between the CreatedDate and ClosedDate of each item.
    iteration_path='IterationPath', # GroupBy.
    new='LeadTimeDays', # New column name.
    literal_filter=['ClosedDate >= 90'], # Filtering items from the last 90 days.
    filters={'WorkItemType': 'Task', 'State': 'Closed'} # Custom filters for metric.
).df

Author

The Azure-Devops-PySpark library was written by Guilherme Silva <https://www.linkedin.com/in/gusantosdev/> in 2022.

https://github.com/gusantos1/Azure-Devops-Spark

License

GNU General Public License v3.0.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

azure-devops-pyspark-1.0.5.tar.gz (24.1 kB view details)

Uploaded Source

File details

Details for the file azure-devops-pyspark-1.0.5.tar.gz.

File metadata

  • Download URL: azure-devops-pyspark-1.0.5.tar.gz
  • Upload date:
  • Size: 24.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.7.9

File hashes

Hashes for azure-devops-pyspark-1.0.5.tar.gz
Algorithm Hash digest
SHA256 24f27a4ba5ae0a60bb9ec4aa02c84c8cfa60b38c5ee33b24b6bdf0d77e454537
MD5 5fb35b0e68c17e0f01625495155c2948
BLAKE2b-256 8a613011069d27f95ef5d25da1e789984c65154be12a6770aa368472e93d5c74

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page