Skip to main content

Software and instructions for setting up and running an autonomous (self-driving) laboratory optics demo using dimmable RGB LEDs, a 10-channel spectrometer, a microcontroller, and an adaptive design algorithm.

Project description

Project generated with PyScaffold PyPI-Server Monthly Downloads Coveralls

Check out the notebooks for some tutorials and demonstrations! See also the manuscript published in Matter [postprint], the build instructions published in Star Protocols, the hackaday page, and the WIP liquid spectrophotometry demo.

self-driving-lab-demo Open In Colab

Software and instructions for setting up and running an autonomous (self-driving) laboratory optics demo using dimmable RGB LEDs, an 8-channel spectrophotometer, a microcontroller, and an adaptive design algorithm, as well as extensions to liquid- and solid-based color matching demos.

Overview

This repository covers three teaching and prototyping demos for self-driving laboratories in the fields of optics (light-mixing), chemistry (liquid-mixing), and solid-state materials science (solid-mixing).

Light-mixing

Self-driving labs are the future; however, the capital and expertise required can be daunting. We introduce the idea of an experimental optimization task for less than $100, a square foot of desk space, and an hour of total setup time from the shopping cart to the first "autonomous drive." For our first demo, we use optics rather than chemistry; after all, light is easier to move than matter. While not strictly materials-based, importantly, several core principles of a self-driving materials discovery lab are retained in this cross-domain example: sending commands to hardware to adjust physical parameters, receiving measured objective properties, decision-making via active learning, and utilizing cloud-based simulations. The demo is accessible, extensible, modular, and repeatable, making it an ideal candidate for both low-cost experimental adaptive design prototyping and learning the principles of self-driving laboratories in a low-risk setting.

Liquid-mixing

We extend the light-mixing demo to a color-matching materials optimization problem using dilute colored dyes. This optimization task costs less than 300 USD, requires less than three square feet of desk space, and less than three hours of total setup time from the shopping cart to the first “autonomous drive.” The demo is modular and extensible; additional peristaltic pump channels can be added, the dye reservoirs can be increased, and chemically sensitive parts can be replaced with chemically resistant ones.

Solid-mixing

There are few to no examples of a low-cost demo platform involving the handling of solid-state materials (i.e., powders, pellets). For this demo, we propose using red, yellow, and blue powdered wax as a replacement for the liquid colored dyes. The demo is more expensive due to the need for robotics. The demo involves using tealight candle holders, transferring them to a rotating stage via a robotic arm, dispensing a combination of powders, melting the wax via an incandescent light bulb, measuring a discrete color spectrum, and moving the completed sample to a separate sample storage area.

Current Users

(listed with permission)

Total verified users: 12

University instructors utilizing CLSLab-Light during Spring 2023: 4 (~40 kits in total)

GroupGets round 1 campaign: funded and fulfilled (19 kits)

GroupGets round 2 campaign: running (current 20 kits)

It is now stocked in the GroupGets Store, albeit with a higher GroupGets fee. If you don't want to wait for new rounds, and you'd rather order a pre-built kit, this is the best option right now.

Hardware

Since the Pico WH is difficult to source (2022-09-15), you will likely need to purchase a Pico W with presoldered headers or purchase a Pico W (see Raspberry Pi Official Resellers) and solder male header pins onto it.

Hardware parts

Assembled

assembled SDL-Demo

Assembled (annotated)

assembled SDL-Demo (annotated)

See Also

Most of the build instructions will go into the Hackaday project page, probably with periodic updates to GitHub. GitHub will host the software that I develop.

The BOM uses Raspberry Pi (RPi) in favor of Arduino to support running complex adaptive design algorithms locally using the higher-end RPi models such as 4B with 8 GB RAM. RPi Zero 2 W and RPi 4B are standalone computers, whereas Arduino typically only has microcontrollers.

Due to the chip shortage, the current setup (2022-08-16) is designed for RPi Pico W (see https://github.com/sparks-baird/self-driving-lab-demo/issues/8), but can be adapted to other models.

ToDo:

Quick Start

https://youtu.be/GVdfJCsQ8vk

Purchase the hardware

  1. See the "DigiKey Order" link along the left column of the Hackaday Page for a standalone order or the Pico W Bill of Materials Wishlist on Adafruit
  2. As of 2023-03-04, PiShop sells a Pico W with presoldered headers, which can circumvent the need for a soldering kit. See also the list of official suppliers. If you purchase a Pico W with separate headers, then solder the headers yourself (or ask a friend or go to a local makerspace or similar).
  3. 14 gauge (outer thickness including insulation) sculpting wire (preferably insulated), e.g. 20 gauge electrical wire at DigiKey or 14 gauge sculpting wire at Amazon

Pico W setup

  1. If you're going to do any soldering, test to make sure that your Pico W works prior to soldering by connecting the USB-A to micro-USB-B cable between your computer and Pico W while holding the BOOTSEL button the Pico. It should open up a notification for a new drive on your computer.
  2. Solder your headers to the Pico if necessary
  3. Plug the Pico W into the Maker Pi Pico base (the board has an image of where the USB port should be pointing to help you know which orientation to put it in)
  4. Cut (or bend repeatedly until it snaps) about 2.5-3 feet of sculpting wire. Loop the same wire through each of the four mounting holes on the Maker Pi Pico base and each of the four mounting holes on the AS7341, and position the AS7341 pinhole sensor perpendicular to and 2-3 inches away from the NeoPixel LED (labeled as GP28)
  5. (Re)connect the USB cable from the Pico W to your computer while holding the BOOTSEL button on the Pico W, and drag-drop the appropriate MicroPython firmware onto the drive that opens up. This will install MicroPython
  6. Download Thonny (for experienced users, you should still download Thonny for initial testing since it has nice features specific to microcontrollers, but of course for the actual programming you can use your preferred IDE)
  7. Download the code from the appropriate src folder (recommended: public_mqtt_sdl_demo)
  8. Rename sample_secrets.py to secrets.py and populate with the necessary WiFi info (and Pico ID if applicable). SSID is basically the WiFi network name. This may not work for school and work networks, so you can use a hotspot instead or only use the nonwireless functionality (nonwireless)
  9. Upload the code to the Pico W
  10. Open main.py and click "run" in Thonny

Running the optimizations

  1. If you choose the public_mqtt_sdl_demo, you can control your SDL-Demo remotely from anywhere. The Jupyter notebook tutorial for this is found at https://github.com/sparks-baird/self-driving-lab-demo/blob/main/notebooks/4.2-paho-mqtt-colab-sdl-demo-test.ipynb.

Advanced Installation

PyPI

conda create -n self-driving-lab-demo python=3.10.*
conda activate self-driving-lab-demo
pip install self-driving-lab-demo

Local

In order to set up the necessary environment:

  1. review and uncomment what you need in environment.yml and create an environment self-driving-lab-demo with the help of conda:
    conda env create -f environment.yml
    
  2. activate the new environment with:
    conda activate self-driving-lab-demo
    

NOTE: The conda environment will have self-driving-lab-demo installed in editable mode. Some changes, e.g. in setup.cfg, might require you to run pip install -e . again.

Optional and needed only once after git clone:

  1. install several pre-commit git hooks with:

    pre-commit install
    # You might also want to run `pre-commit autoupdate`
    

    and checkout the configuration under .pre-commit-config.yaml. The -n, --no-verify flag of git commit can be used to deactivate pre-commit hooks temporarily.

  2. install nbstripout git hooks to remove the output cells of committed notebooks with:

    nbstripout --install --attributes notebooks/.gitattributes
    

    This is useful to avoid large diffs due to plots in your notebooks. A simple nbstripout --uninstall will revert these changes.

Then take a look into the scripts and notebooks folders.

Dependency Management & Reproducibility

  1. Always keep your abstract (unpinned) dependencies updated in environment.yml and eventually in setup.cfg if you want to ship and install your package via pip later on.
  2. Create concrete dependencies as environment.lock.yml for the exact reproduction of your environment with:
    conda env export -n self-driving-lab-demo -f environment.lock.yml
    
    For multi-OS development, consider using --no-builds during the export.
  3. Update your current environment with respect to a new environment.lock.yml using:
    conda env update -f environment.lock.yml --prune
    

Other Builds

Closed-loop Spectroscopy Laboratory: Liquid

CLSLab is an autonomous self-driving laboratory that uses dilluted food dye, peristaltic pumps, a light source, and a light sensor to perform liquid-based color-matching to a target color. See the following visual summary:

See https://github.com/sparks-baird/self-driving-lab-demo/discussions/89 for discussion of ongoing development.

Bill of Materials

  • Everything from the CLSLab:Light demo
    • Pico W + headers (need to solder headers)
    • Maker Pi Pico base
    • Stemma-QT to Grove connector
    • AS7341 Light Sensor
    • USB-A to micro-USB-B cable
    • 5V USB-A wall adapter
    • 18 AWG sculpting wire
  • Plant drip bag (large) x2
  • Party drink drip bag (small) x3
  • Spectrophotometer cuvette
  • Peristaltic pump x5
  • silicone (or PTFE?) tubing (3 m) (what inner and outer diameter?)
  • LED light source (see https://github.com/sparks-baird/self-driving-lab-demo/discussions/52 for context)
  • tube connectors (which kind?)
  • double Wye connectors
  • silicone-to-glass glue or epoxy (to permanently fix a silicone lid to a cuvette)

The plant and party IV-style drip bags can be replaced with medical-grade chemically resistant versions that are made up of e.g. polyethylene.

Required Tools, etc.

Project Organization

├── AUTHORS.md              <- List of developers and maintainers.
├── CHANGELOG.md            <- Changelog to keep track of new features and fixes.
├── CONTRIBUTING.md         <- Guidelines for contributing to this project.
├── Dockerfile              <- Build a docker container with `docker build .`.
├── LICENSE.txt             <- License as chosen on the command-line.
├── README.md               <- The top-level README for developers.
├── configs                 <- Directory for configurations of model & application.
├── data
│   ├── external            <- Data from third party sources.
│   ├── interim             <- Intermediate data that has been transformed.
│   ├── processed           <- The final, canonical data sets for modeling.
│   └── raw                 <- The original, immutable data dump.
├── docs                    <- Directory for Sphinx documentation in rst or md.
├── environment.yml         <- The conda environment file for reproducibility.
├── models                  <- Trained and serialized models, model predictions,
│                              or model summaries.
├── notebooks               <- Jupyter notebooks. Naming convention is a number (for
│                              ordering), the creator's initials and a description,
│                              e.g. `1.0-fw-initial-data-exploration`.
├── pyproject.toml          <- Build configuration. Don't change! Use `pip install -e .`
│                              to install for development or to build `tox -e build`.
├── references              <- Data dictionaries, manuals, and all other materials.
├── reports                 <- Generated analysis as HTML, PDF, LaTeX, etc.
│   └── figures             <- Generated plots and figures for reports.
├── scripts                 <- Analysis and production scripts which import the
│                              actual PYTHON_PKG, e.g. train_model.
├── setup.cfg               <- Declarative configuration of your project.
├── setup.py                <- [DEPRECATED] Use `python setup.py develop` to install for
│                              development or `python setup.py bdist_wheel` to build.
├── src
│   └── self_driving_lab_demo <- Actual Python package where the main functionality goes.
├── tests                   <- Unit tests which can be run with `pytest`.
├── .coveragerc             <- Configuration for coverage reports of unit tests.
├── .isort.cfg              <- Configuration for git hook that sorts imports.
└── .pre-commit-config.yaml <- Configuration of pre-commit git hooks.

Note

This project has been set up using PyScaffold 4.2.3.post1.dev10+g7a0f254 and the dsproject extension 0.7.2.post1.dev3+g948a662.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

self_driving_lab_demo-0.8.2-py3-none-any.whl (1.1 MB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page