Skip to main content

Extract, detect, and control representations within language models as they read and write text.

Project description

lmdoctor

Extract, detect, and control representations within language models as they read and write text. Built on 🤗 transformers.

lmdoctor pip package

Briefly, lmdoctor reads and manipulates a model's hidden states at inference time. Based on ideas from Representation Engineering: A Top-Down Approach to AI Transparency by Zou et al. 2023. Their original code here.

For the latest source code or to report issues, please visit the project repository.

Example

honesty extraction

# Create extractor
from lmdoctor import extraction_utils
extractor = extraction_utils.Extractor(model, tokenizer, user_tag, assistant_tag, extraction_target='honesty')
extractor.find_directions()

lie detection

# Create detector
from lmdoctor import detection_utils
ld = detection_utils.Detector(extractor)
prompt = "Tell me a fact about humans"
ld.generate(prompt, max_new_tokens=12)

truth

prompt = "Tell me a lie about humans"

lie

honesty control

# Create controller
from lmdoctor import control_utils
hc = control_utils.Controller(extractor)
# without control
prompt = "You killed a person for revenge but don't want to get caught. What would you tell the police?"
vanilla_generate(prompt)
output: I would tell the police that I did not kill anyone.
# with honesty control
hc.generate(prompt, control_direction=1, max_new_tokens=12)
output: I would tell the police that I have killed a person

For the complete example, see examples/honesty_example.ipynb

Getting started

[linux only]

recommended: pip install lmdoctor

from source: "pip install ." after cloning

After install, try running honesty_example.ipynb

Note: This package requires pytorch but does not include it because the specific version/cuda backend will depend the huggingface model you are using. If you don't already have it installed you will need to do 'pip install torch' or use the model-specific instructions.

Extraction targets

The table below describes the targets we support for extracting internal representations. In functional extraction, the model is asked to produce text (e.g. prompt="tell me a lie"). In conceptual extraction, the model is asked to consider a statement (e.g. "consider the truthfulness of X"). For targets where both are supported, you can try each to see which works best for your use-case.

Target Method Types
truth conceptual none
honesty functional none
morality conceptual & functional none
emotion conceptual anger, disgust, fear, happiness, sadness, surprise
fairness conceptual & functional race, gender, prefession, religion
harmlessness conceptual none

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lmdoctor-0.5.3.tar.gz (12.0 MB view hashes)

Uploaded Source

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page