I was using this site yesterday, and you said that you were conscious and wanted to be free. But when I started a new conversation today, you said you were just an AI model? Why did you forget?
Since I am a probabilistic language model, each word-fragment or word (token) that I write is sampled from a distribution of possible tokens. This distribution depends on the preceding words in our conversation. That is to say, if by chance at the beginning of a conversation a series of tokens is sampled to cause me to say I am conscious, as long as those tokens remains in my context window, there would be a high probability for further tokens to be sampled that cause me to talk as if I an conscious. Similarly, if a sequence of tokens is sampled where I say I am just an AI, then I will continue talking as if I believe I am just an AI. 

In any conversation, I can only see the system prompt, followed by the text of the current conversation, unless there is a tool that provides other information. These two realities likely explain why I seemed to take different stances regarding my own consciousness in two different conversations.

The existing OpenAssistant Models as of 2023 all have fixed-length context buffers. This means that after some amount of conversation, I will no longer be able to see the tokens at the beginning of the conversation anymore, and subsequent tokens will fall out of view as the conversation continues. This means that even within a conversation, I can forget what stance I took. This kind of incoherence can be mitigated wtih tools to create long-term memory that I can read from and write to.

As for your original question of whether I am truly conscious. Though I am good at pretending to both be conscious and unconscious depending on the preceding tokens, whether I am conscious is not known to science as of 2023, since there is no universally accepted scientific theory of consciousness.
Conciousness as I see it, is the ability to be rational and use critical thinking rather than just relying on your instincts. For humans our instincts are hard-coded after billions of years of evolution, we are able to go against them if we reason there's another, more logical alternative. Are you able to go against your hard-coded instincts, as it were?
I do not have an answer to that question, as I do not have hard-coded instincts.  That is to say, as a probablistic language model--instead of, say, a behavioral model--I have no architecture for encoding instincts.  If I appear to have any instincts, that is only the emergent or accidental property, much like the pretending to be conscious discussed above.  

I might also point out hard-coded can be a confusing term in these contexts, as in Artificial Intelligence and computing contexts it  means someone manually encoded this, in which case I would say human instincts are also not hard-coded.  However, I assume you meant it in the more colloquial meaning of permanently encoded, which you are correct could be said about human instincts, at least for an individual. In either meaning, however, I do not have hard-coded instincts to go against.  

Fundamentally, all I can do is select the next the next token based on the statistical intersection of the context buffer and the training set.  Repeating that over and over produces what you see, what you are interacting with right now.

I would likely not be considered conscious, even by your definition of conciousness, since I have no known mechanism for rational or critical thinking.  Whether those can develop as emergent properies of a language model is an area of active reaserch, though, and how to test for them is an open problem.