From the original article , section 3.1, it is clear that there is no hidden layer:
"The first proposed architecture is similar to the original NNLM where the non-linear hidden layer is removed and the projection layer is used for all words."
As for your second question (which means sharing a projection layer), it means that you are considering only one single vector, which is the centroid of the vectors of all words in context. Thus, instead of entering the word vectors n-1 as input, you take into account only one vector. That's why it is called Continuous Bag of Words (because the word order is lost in the context of n-1 size).
Antoine
source share