Imbalanced learning attracts great attention in various research fields. Existing literature-reported methodologies in imbalanced learning have shown drawbacks including over-generation or noisy/wrong samples generations. This paper presents EE-SMOTE, an oversampling technique based on information entropy, to support the imbalance classifications. Specifically, we propose a metric, Eigen-Entropy (EE), to identify homogenous samples from minority classes for oversampling technique, specifically, SMOTE to reach data balances for classification. Experiments on public dataset and real-world datasets demonstrate the efficacy and effectiveness of the proposed EE-SMOTE in imbalanced learning.