Compare commits
9 Commits
revert-8-f
...
feature/15
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
41086e95ad | ||
|
|
adde35ed7e | ||
|
|
b2684c23f6 | ||
|
|
8a499a04fb | ||
|
|
118c56c12d | ||
|
|
79a0f82372 | ||
|
|
c9415c21fa | ||
|
|
3860f2cc5b | ||
|
|
553140fe3c |
2
.gitignore
vendored
2
.gitignore
vendored
@@ -1,4 +1,4 @@
|
||||
# Ignore CSV files in the data directory and all its subdirectories
|
||||
data/**/*.csv
|
||||
|
||||
.venv/
|
||||
*.pyc
|
||||
File diff suppressed because one or more lines are too long
@@ -1,8 +1,8 @@
|
||||
# Processed Data Directory
|
||||
# Raw Data Directory
|
||||
|
||||
## Overview
|
||||
|
||||
This `data/processed` directory contains structured data that has been processed and formatted for analysis. Each subdirectory within `processed` represents a different level of simulated damage, and each contains multiple test files from experiments conducted under that specific damage scenario.
|
||||
This `data/raw` directory contains structured data that has been processed and formatted for analysis. Each subdirectory within `raw` represents a different level of simulated damage, and each contains multiple test files from experiments conducted under that specific damage scenario.
|
||||
|
||||
## Directory Structure
|
||||
|
||||
@@ -12,12 +12,12 @@ The directory is organized as follows:
|
||||
data
|
||||
└── processed
|
||||
├── DAMAGE_1
|
||||
│ ├── D1_TEST1.csv
|
||||
│ ├── D1_TEST2.csv
|
||||
│ ├── D1_TEST1.csv
|
||||
│ ├── D1_TEST2.csv
|
||||
│ ...
|
||||
│ └── D1_TEST10.csv
|
||||
│ └── D1_TEST10.csv
|
||||
├── DAMAGE_2
|
||||
│ ├── D2_TEST1.csv
|
||||
│ ├── D2_TEST1.csv
|
||||
│ ...
|
||||
├── DAMAGE_3
|
||||
│ ...
|
||||
@@ -13,29 +13,38 @@ processed_path = os.path.join(base_path, "processed")
|
||||
os.makedirs(raw_path, exist_ok=True)
|
||||
os.makedirs(processed_path, exist_ok=True)
|
||||
|
||||
for damage in range(1, 6): # 5 Damage levels
|
||||
damage_folder = f"DAMAGE_{damage}"
|
||||
damage_path = os.path.join(processed_path, damage_folder)
|
||||
# Define the number of zeros to pad
|
||||
num_damages = 5
|
||||
num_tests = 10
|
||||
num_sensors = 2
|
||||
damage_pad = len(str(num_damages))
|
||||
test_pad = len(str(num_tests))
|
||||
sensor_pad = len(str(num_sensors))
|
||||
|
||||
for damage in range(1, num_damages + 1): # 5 Damage levels starts from 1
|
||||
damage_folder = f"DAMAGE_{damage:0{damage_pad}}"
|
||||
damage_path = os.path.join(raw_path, damage_folder)
|
||||
os.makedirs(damage_path, exist_ok=True)
|
||||
|
||||
for test in range(1, 11): # 10 Tests per damage level
|
||||
# Filename for the CSV
|
||||
csv_filename = f"D{damage}_TEST{test}.csv"
|
||||
csv_path = os.path.join(damage_path, csv_filename)
|
||||
for sensor in range(1, 3): # 2 Sensors per test
|
||||
# Filename for the CSV
|
||||
csv_filename = f"D{damage:0{damage_pad}}_TEST{test:0{test_pad}}_{sensor:0{sensor_pad}}.csv"
|
||||
csv_path = os.path.join(damage_path, csv_filename)
|
||||
|
||||
# Generate dummy data
|
||||
num_rows = 10
|
||||
start_time = datetime.now()
|
||||
timestamps = [start_time + timedelta(seconds=i*0.0078125) for i in range(num_rows)]
|
||||
values = np.random.randn(num_rows) # Random float values
|
||||
|
||||
# Generate dummy data
|
||||
num_rows = 10
|
||||
start_time = datetime.now()
|
||||
timestamps = [start_time + timedelta(seconds=i*0.0078125) for i in range(num_rows)]
|
||||
values = np.random.randn(num_rows) # Random float values
|
||||
# Create DataFrame
|
||||
df = pd.DataFrame({
|
||||
"Time": timestamps,
|
||||
"Value": values
|
||||
})
|
||||
|
||||
# Create DataFrame
|
||||
df = pd.DataFrame({
|
||||
"Time": timestamps,
|
||||
"Value": values
|
||||
})
|
||||
|
||||
# Save the CSV file with a custom header
|
||||
with open(csv_path, 'w') as file:
|
||||
file.write('sep=,\n') # Writing the separator hint
|
||||
df.to_csv(file, index=False)
|
||||
# Save the CSV file with a custom header
|
||||
with open(csv_path, 'w') as file:
|
||||
file.write('sep=,\n') # Writing the separator hint
|
||||
df.to_csv(file, index=False)
|
||||
|
||||
Reference in New Issue
Block a user