Q-learning is a kind of reinforcement learning where the agent solves the given task based on rewards received from the environment. The most of researches done in the field of reinforcement learning has focused on the discrete domains. But the environment with which the agent should interact is continuous. Thus a method that is able to make Q-learning applicable to the continuous problem domain is needed. In this thesis, basic fuzzy rule is extended so that it can incorporate the Q-learning. And the interpolation technique which is widely used in memory-based learning is adopted to represent the appropriate Q value for current state and action pair. The resulting structure based on fuzzy inference system has the capability of solving the continuous state and action problem in Q-learning by virtue of fuzzy inference system. In addition, the resulting Self-Organizing Fuzzy Inference System by Q-learning(SOFIS-Q) can generate fuzzy rules via interacting with the environment without a priori knowledge about the environment. The effectiveness of proposed structure is shown thorough simulation on cart-pole system.