-Theinclusionoffakedatawascarefullyevaluatedasitsignificantlyinfluencedthemodel's accuracy. Adding more than 25% fake data caused the accuracy to drop drastically, making the model unreliable.
- To maintain balance and robustness, the fake data was limited to 25%. This ensured enough variety without heavily compromising model performance.
- The evaluation showed a decline in performance metrics such as **Mean Squared Error (MSE)** and **R²** scores when fake data exceeded this threshold, emphasizing the importance of minimizing synthetic data.
### 6. Data Splitting
- The dataset was split into **training** and **testing** sets using a **70:30** ratio for model evaluation.