This project was developed as part of the Green-AI Hub pilot initiative in collaboration with the company Heismann. The aim was to optimize internal processes through AI-driven solutions.
Two main use cases were addressed:
-
Machine Warm-up Time Analysis: Statistical analysis was performed to determine the optimal start temperature of machines based on part production data and temperature measurements.
-
Part Dimension Prediction: Machine learning models were trained to predict the deviation of part dimensions from optimal values, allowing early detection of defective parts with significant deviations.
To protect company-sensitive data, some scripts and the database were excluded from publication. Additional anonymization steps were performed, so some scripts may be incomplete or placeholder-based.
Since no final prototype was developed, the project has been organized into logical folders representing individual work steps:
-
Database: PostgreSQL-based database. Scripts for loading and reading data are provided. To use it, start a PostgreSQL instance and either restore a backup using the provided batch file or configure a fresh one via
config.json
.Note: The actual database is not publicly included. It contains all project-related data, including:
- Measurement data of the components
- Temperature data
- Vibration data
- Machine energy data
- Axis settings
- Tool changes
- Production times
- Rotational speeds
- ...
-
Dataparser: All raw data (TXT and CSV) was collected from machines and provided by the company. These files were parsed and imported into the database using specialized scripts—each file type required its own parser due to inconsistent formats.
-
Model_Training: Models were trained on a computing cluster using TPOT AutoML. Various preprocessing steps were applied to prepare optimal training data. Due to their high frequency, temperature and frequency data were not loaded into the database but joined directly from the original files during training.
-
Visualisations: Contains all scripts used to generate visual representations of the data. Each file describes the purpose of the specific visualization.
-
Experimental: Contains an unfinished third use case: a statistical analysis of tool lifespan. Some initial scripts for formula definition and visual evaluation are included. This area was not further developed.
Clone this repository, open a terminal, and navigate to the project directory.
To install the required Python dependencies, run:
pip install -r requirements.txt
To set up your own PostgreSQL instance, install PostgreSQL via this link and configure your database as described in the Database folder.
Note: Currently, the scripts are tested only with Python 3.8.
Contributions are what make the open-source community such a great place to learn, inspire, and grow. Any contributions you make are greatly appreciated.
If you have a suggestion for improvement, please fork the repository and submit a pull request. You can also open an issue with the tag "enhancement"
.
Steps to contribute:
- Fork the Project
- Create your Feature Branch (
git checkout -b feature/AmazingFeature
) - Commit your Changes (
git commit -m 'Add some AmazingFeature'
) - Push to the Branch (
git push origin feature/AmazingFeature
) - Open a Pull Request
Distributed under the MIT License. See LICENSE.txt
for details.
Green-AI Hub Mittelstand – [email protected] Project Link: https://github.com/Green-AI-Hub-Mittelstand/repository_name
Get in touch »