In human-centric technologies, skin segmentation of body parts is a prerequisite for high-level processing. The traditional method of skin detection is pixel-wise detection coupled with morphological operations. Pixel-wise methods usually generate a number of false samples and outlier skin pixels, which can make it difficult for morphological operations to provide satisfactory results in complex scenarios. Furthermore, in many cases only a coarse region is required (e.g., the bounding-box of the face) rather than detailed pixel-wise labeling. A patch-wise skin segmentation method is proposed based on deep neural networks. Our method treats image patches as processing units instead of pixels, which directly exploits the spatial information of pixels in the detection stage rather than using morphological operations on isolated pixels after detection. An image patch dataset is built and deep skin models (DSMs) are trained based on the new dataset. Trained DSMs are then integrated into a sliding window framework to segment skin regions of the human body parts. Experiments on standard benchmarks demonstrate that DSMs provide more explicit skin region of interest candidates than pixel-wise methods in complex scenarios, and achieve competitive performance on pixel-wise skin detection.