Image denoising approaches that learn spatially adaptive dictionaries from the observed noisy image have gathered a lot of attention in the past decade. These methods rely on the hypothesis that patches from the underlying clean image can be expressed as sparse linear combinations of these dictionary vectors (bases). We present a framework for inferring an orthonormal set of dictionary vectors using orthogonal locality preserving projection (OLPP). This ensures that patches that are similar in the noisy image should produce similar coefficients when projected in the OLPP domain. Unlike other projection methods, the locality preserving property of OLPP automatically groups similar patches together during inference of the basis. Hence, only one global orthonormal basis suffices to sparsely represent patches from a large subimage or a large portion of the image. The proposed amalgamation of the sparsity and global dictionary make the current approach more suitable for an image denoising task with reduced computational complexity. Experiments on several benchmark datasets made it clear that the proposed method is capable of preserving fine textures while denoising an image, on par with or surpassing several state-of-the-art methods for gray-scale and color images.