Skip to main content

protein turnover pipeline

Project description

Protein Turnover

protein turnover computations

Install with:

pip install protein-turnover
# If you plan to use non-western text for your job names maybe also install unidecode
# pip install Unidecode
# *OR*
pip install protein-turnover[unidecode]

This will give you a turnover command (equivalent to python -m protein_turnover).

The Turnover Job File

To run protein-turnover you need to create a jobfile (which is in toml format).

e.g.:

# job name is a display name and should contain information about what the job is about.
job_name = "My Experiment"
pepxml = "chx_cc_repeat.interact.pep.xml"
protxml = "chx_cc_repeat.combined.prot.xml"
# a list of mzML files
mzmlfiles = [ "milla009642.mzML"]
# internal job identifier (*optional* used to create auxilary filenames)
jobid = "job1"
# for cached data. If not specifies cache files will be placed in the
# same directories as original datafiles
cache_dir = "."
# email is *optional*
email = "me.lastname@uwa.edu.au"

[settings]
# these are the default settings
rtTolerance = 15.0
mzTolerance = 1e-5
labelledIsotopeNumber = 15
labelledElement = "N"
maximumLabelEnrichment = 0.95
retentionTimeCorrection = "SimpleMedian"
useObservedMz = false
minProbabilityCutoff = 0.8
enrichmentColumns = 10

So a minimal jobfile would be (say):

job_name = "My Experiment"
pepxml = "chx_cc_repeat.interact.pep.xml"
protxml = "chx_cc_repeat.combined.prot.xml"
mzmlfiles = [ "milla009642.mzML"]

Notes:

  • email will only work if the config.MAIL_SERVER is correct.
  • job_name is really just a human readable short description of the job.
  • jobid is used (mainly) to create filenames; for example the final sqlite output file will be called {jobid}.sqlite
  • File names that are not absolute are relative to the current working directory of the turnover process.
  • If [settings] is missing the values will default to the example values above. You only need to specify values that are different from the ones above.
  • cache_dir: see below.

Running a Job

turnover run {jobfile}.toml
# *OR*
python -m protein_turnover run {jobfile}.toml

# alter configuration and use info level logging and log to logfile.log
turnover --level=info --logfile=logfile.log run {jobfile}.toml

Cache Files and cache_dir

Turnover translates all the .mzML, pep.xml, and prot.xml files into pandas DataFrames stored in .parquet format, plus an internal (to turnover) format that make it easy to quickly scan spectra using mmap.

These files are cached in cache_dir based on an sha256 hash of the contents of each file. Thus re-runs of the job don't need to (re)-generate these files again.

Because of the sha256 hash you can used a single cache_dir for all jobs.

If the cache files are deleted, they will be recreated when the job is run again.

If cache_dir is not specified the the cached files will be placed in the same directory as the originator xml files.

Viewing

One the job has run you can view the results in a browser

pip install protein-turnover-website
turnover view {jobfile}.toml

Windows

A default install of python on windows Will give you a py function instead of a python function. Go to the search bar and type cmd. In the cmd shell you should use instead of turnover ... py -m protein_turnver ...

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

protein_turnover-0.5.8.tar.gz (76.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

protein_turnover-0.5.8-py3-none-any.whl (92.8 kB view details)

Uploaded Python 3

File details

Details for the file protein_turnover-0.5.8.tar.gz.

File metadata

  • Download URL: protein_turnover-0.5.8.tar.gz
  • Upload date:
  • Size: 76.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.1.1 CPython/3.12.3 Linux/6.11.0-17-generic

File hashes

Hashes for protein_turnover-0.5.8.tar.gz
Algorithm Hash digest
SHA256 1b7b209b8c17d6ca5390541c13ad472c37749446f15aa653fbfbd1ccfcd2473e
MD5 f9ad91f36911234379abd28d06ad2d4b
BLAKE2b-256 d4f430de25b250ecdb4388b00351c252e62d335ba1747a742da047eb77aea4f7

See more details on using hashes here.

File details

Details for the file protein_turnover-0.5.8-py3-none-any.whl.

File metadata

  • Download URL: protein_turnover-0.5.8-py3-none-any.whl
  • Upload date:
  • Size: 92.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.1.1 CPython/3.12.3 Linux/6.11.0-17-generic

File hashes

Hashes for protein_turnover-0.5.8-py3-none-any.whl
Algorithm Hash digest
SHA256 8d09c9f6a5bccfa41fcc7affcec5e112ee3f76e4a7ddd5b865d564d4beb2b8a5
MD5 2d70cca5b21fb90309bca0b3724efcf5
BLAKE2b-256 f89ce2750174f9e118c4c1882151204b3f210a15ca5264688eb823750dc8c50e

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page