With the depressive psychiatric disorders becoming more common, people are gradually starting to take it seriously. Somatisation disorders, as a general mental disorder, are rarely accurately identified in clinical diagnosis for its specific nature. In the previous work, speech recognition technology has been successfully applied to the task of identifying somatisation disorders on the Shenzhen Somatisation Speech Corpus. Nevertheless, there is still a scarcity of labels for somatisation disorder speech database. The current mainstream approaches in the speech recognition heavily rely on the well labelled data. Compared to supervised learning, self-supervised learning is able to achieve the same or even better recognition results while reducing the reliance on labelled samples. Moreover, self-supervised learning can generate general representations without the need for human hand-crafted features depending on the different recognition tasks. To this end, we apply self-supervised learning pre-trained models to solve few-labelled somatisation disorder speech recognition. In this study, we compare and analyse the results of three self-supervised learning models (contrastive predictive coding, wav2vec and wav2vec 2.0). The best result of wav2vec 2.0 model achieves 77.0 % unweighted average recall and is significantly better than CPC (p < .005), performing better than the benchmark of the supervised learning model.Clinical relevance— This work proposed a self-supervised learning model to resolve the few-labelled SD speech data, which can be well used for helping psychiatrists with clinical assistant to diagnosis. With this model, psychiatrists no longer need to spend a lot of time labelling SD speech data.