feat(data): Initialize dummy data

- Create a Python script to generate CSV files in a structured folder hierarchy under `data/processed` with specific damage levels and tests.
- Add a `.gitignore` file to exclude CSV files from Git tracking, enhancing data privacy and reducing repository size.
- Include a `README.md` in the `data` directory to clearly document the directory structure, file content, and their intended use for clarity and better usability.

Closes #7
This commit is contained in:
nuluh
2024-08-14 18:15:08 +07:00
parent feb3c85340
commit 153e8cb109
3 changed files with 99 additions and 0 deletions

56
data/processed/README.md Normal file
View File

@@ -0,0 +1,56 @@
# Processed Data Directory
## Overview
This `data/processed` directory contains structured data that has been processed and formatted for analysis. Each subdirectory within `processed` represents a different level of simulated damage, and each contains multiple test files from experiments conducted under that specific damage scenario.
## Directory Structure
The directory is organized as follows:
```
data
└── processed
├── DAMAGE_1
│ ├── D1_TEST1.csv
│ ├── D1_TEST2.csv
│ ...
│ └── D1_TEST10.csv
├── DAMAGE_2
│ ├── D2_TEST1.csv
│ ...
├── DAMAGE_3
│ ...
├── DAMAGE_4
│ ...
└── DAMAGE_5
```
Each subdirectory (`DAMAGE_1`, `DAMAGE_2`, etc.) represents a specific damage scenario. Inside each damage directory, there are CSV files named in the format `Dx_TESTy.csv` where `x` is the damage level and `y` is the test number. There are 10 tests for each damage level.
## File Description
Each `.csv` file contains the following columns:
- **Time**: Timestamp of the measurement in the format `MM/DD/YY HH:MM:SS.sssssssss`.
- **Value**: A float value representing the measurement taken at the corresponding timestamp.
The files begin with a custom header indicating the separator used (`sep=,`), ensuring compatibility with various CSV parsers that support separator hints.
## Purpose
The processed data is intended for use in analytical models that assess structural integrity under various simulated damage scenarios. The tests vary by parameters such as the intensity and type of applied stress, allowing for comprehensive analysis across different conditions.
## Usage
To use these files in your analysis:
1. Ensure your CSV parser can interpret the custom header.
2. Load the data from the desired `Dx_TESTy.csv` file.
3. Analyze the data according to your specific needs (e.g., regression analysis, classification).
The data can be directly imported into most data analysis software packages or programming environments that support CSV file input.
## Note
Please do not manually edit the CSV files to ensure data integrity. If processing adjustments are necessary, consider scripting the changes and documenting any transformations applied.