Proper grouping in classifier algorithms is a critical element of supervised learning. The first step in this is to have the correct data. Data that has a problem is worse than not having it. One of the biggest problems inherent in natural data is its imbalance. For a classifier algorithm to achieve its best performance, the first step is to fix the problem of data imbalance. To work with real datasets, the first step is to balance the data. The main problem with existing algorithms is to duplicate minority data and generate data that make outlines part of the primary data. The Statistic Deviation Mode Balancer (SDMB) algorithm solves this problem by making samples that adhere to the original data structure. Our proposed algorithm generates data that is very similar to the original data with the help of standard deviation and the amount of minor data mode and moving away from the majority part. Using these two parameters, the SDMB algorithm avoids Outlier data and generates clean data. The output of this algorithm is a balance datasheet that helps classifier algorithms learn the best way from the data. Different classifier algorithms with entirely different methods have been tested to prove this point. First, we balanced the different datasets with our method. Then, with varying classifier algorithms, we compared it with other existing algorithms. This experiment showed that our proposed algorithm is superior to other competitors and can be used in the work process of real datasets.