Determining the optimal number of clusters in cluster analysis without prior knowledge remains a critical and challenging task. Existing methods often depend on calculating clustering validity indices (CVIs), which increases complexity and may reduce efficiency. Furthermore, different CVIs frequently suggest varying optimal cluster numbers, complicating the selection process. To address these challenges, we propose a novel clustering algorithm, self-regulating possibilistic C-means (PCM) with high-density points (SR-PCM-HDP), which simplifies cluster number determination while improving clustering efficiency. First, the density-based knowledge extraction (DBKE) method is introduced to estimate an appropriate initial cluster number and identify high-density points. DBKE enhances the density peak clustering (DPC) algorithm by removing the need for a predefined density radius. Second, SR-PCM-HDP refines the clustering process by incorporating a parameter to balance the interactions between high-density points and cluster centers, reducing sensitivity to initial configurations and accelerating convergence. Third, the parameter adjustment mechanism in classical PCM is redefined to enable adaptive updates during SR-PCM-HDP iterations. This mechanism facilitates the gradual elimination of obsolete clusters and iterative cluster formation. The theoretical foundations of the SR-PCM-HDP cluster elimination mechanism are rigorously established. Experimental results validate the accuracy and effectiveness of SR-PCM-HDP in determining cluster numbers and ensuring clustering validity, particularly for datasets with overlapping or imbalanced distributions. Comparisons are conducted against 13 state-of-the-art algorithms, including fuzzy clustering, possibilistic clustering, and CVI-based cluster determination methods.