Recently, the Q-Learning algorithm, which is one kind of reinforcement learning, is mainly used to implement artificial intelligence system in combination with deep learning. Many research is going on to improve the performance of Q-Learning. Therefore, purpose of theory try to improve the performance of Q-Learning algorithm. This Theory apply Cross Entropy Error to the loss function of Q-Learning algorithm. Since the mean squared error used in Q-Learning is difficult to measure the exact error rate, the Cross Entropy Error, known to be highly accurate, is applied to the loss function. Experimental results show that the success rate of the Mean Squared Error used in the existing reinforcement learning was about 12% and the Cross Entropy Error used in the deep learning was about 36%. The success rate was shown.
This study is the first attempt to raise face recognition ability through machine learning algorithm and apply to CRM's information gathering, analysis and application. In other words, through face recognition of VIP customer in distribution field, we can proceed more prompt and subdivided customized services. The interest in machine learning, which is used to implement artificial intelligence, has increased, and it has become an age to automate it by using machine learning beyond the way that a person directly models an object recognition process. Among them, Deep Learning is evaluated as an advanced technology that shows amazing performance in various fields, and is applied to various fields of image recognition. Face recognition, which is widely used in real life, has been developed to recognize criminals' faces and catch criminals. In this study, two image analysis models, TF-SLIM and Inception-V3, which are likely to be used for criminal face recognition, were selected, analyzed, and implemented. As an evaluation criterion, the image recognition model was evaluated based on the accuracy of the face recognition program which is already being commercialized. In this experiment, it was evaluated that the recognition accuracy was good when the accuracy of the image classification was more than 90%. A limit of our study which is a way to raise face recognition is left as a further research subjects.
Recently, there are many companies that use systems based on artificial intelligence. The accuracy of artificial intelligence depends on the amount of learning data and the appropriate algorithm. However, it is not easy to obtain learning data with a large number of entity. Less data set have large generalization errors due to overfitting. In order to minimize this generalization error, this study proposed DGA which can expect relatively high accuracy even though data with a less data set is applied to machine learning based genetic algorithm to deep learning based dropout. The idea of this paper is to determine the active state of the nodes. Using Gradient about loss function, A new fitness function is defined. Proposed Algorithm DGA is supplementing stochastic inconsistency about Dropout. Also DGA solved problem by the complexity of the fitness function and expression range of the model about Genetic Algorithm As a result of experiments using MNIST data proposed algorithm accuracy is 75.3%. Using only Dropout algorithm accuracy is 41.4%. It is shown that DGA is better than using only dropout.
The range of problems that can be handled by the activation of big data and the development of hardware has been rapidly expanded and machine learning such as deep learning has become a very versatile technology. In this paper, mnist data set is used as experimental data, and the Cross Entropy function is used as a loss model for evaluating the efficiency of machine learning, and the value of the loss function in the steepest descent method is We applied the Gradient Descent Optimize algorithm to minimize and updated weight and bias via backpropagation. In this way we analyze optimal reliability value corresponding to the number of exercises and optimal reliability value without overfitting. And comparing the overfitting time according to the number of data changes based on the number of training times, when the training frequency was 1110 times, we obtained the result of 92%, which is the optimal reliability value without overfitting.