mft to parquet (pyarrow dtypes)
Project description
mft to parquet (pyarrow dtypes)
Tested against Windows 10 / Python 3.11 / Anaconda
pip install mft2parquet
Reads HDD (Hard Disk Drive) information from a specified drive and returns it as a pandas DataFrame.
Args:
drive (str, optional): The drive path to read from. Default is "c:\\".
outputfile (str, optional): If provided, the DataFrame will be saved as a Parquet file at this path.
Default is None.
Returns:
pd.DataFrame: A DataFrame with pyarrow dtypes containing HDD information with the specified columns.
Raises:
subprocess.CalledProcessError: If the external command fails to execute.
Note:
- This function uses an external command-line utility https://github.com/githubrobbi/Ultra-Fast-File-Search to retrieve HDD information.
- The DataFrame will have the following columns:
- aa_path
- aa_name
- aa_path_only
- aa_size
- aa_size_on_disk
- aa_created
- aa_last_written
- aa_last_accessed
- aa_descendents
- aa_read-only
- aa_archive
- aa_system
- aa_hidden
- aa_offline
- aa_not_content_indexed_file
- aa_no_scrub_file
- aa_integrity
- aa_pinned
- aa_unpinned
- aa_directory_flag
- aa_compressed
- aa_encrypted
- aa_sparse
- aa_reparse
- aa_attributes
Example:
df = read_hdd(drive="d:\\", outputfile="hdd_info.parquet")
# Reads HDD information from the 'D:' drive and saves it as 'hdd_info.parquet'.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
mft2parquet-0.10.tar.gz
(5.6 kB
view hashes)
Built Distribution
Close
Hashes for mft2parquet-0.10-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | b3a8fe52a740a327c1418ae2166012ebab010abed8318f11b65f74182bfd7b66 |
|
MD5 | c2ea8d0fb02d385a67569d8f1511d673 |
|
BLAKE2b-256 | 6122c823856c9ae5f6618ffe0bd388c3b6dd055faf3b7b544f14e94fbae8046f |