I would like to thank everyone of you who supported me during this 3 months journey. Your tips gave me the strength that helped me to create better content and write better code. If you are interested in more details why I feel like a winner (and why you should do) – read more. Thank you!
This is my last technical post in the Get Noticed 2017 category. The last one will be a quick summary of the last three months. Soo…it was a very long week. I didn’t do too much but there was a big progress in using the neural network in the Aksesi Proxy Application.
The neural network requires to train it before using. We are expected to provide sets od date that will be used for the purpose of learning. It means that we need a generator which will create gestures in a form readable by the NN.
Today we are taking part in Rzeszów Hackathon. Our team consists of 4 members – Bartek, Vlad, Michał and me. By the way, you should know Bartek and Michał from the post Pekaton – 24-godzinny hackaton. At the beginning, I want to mention that in this post I’m not going to reveal project’s main idea and implementation details. This is because some of ours competitors may read it (good luck guys ;)). Maybe we will share a few hints. Since the start, this post will be updated every hour, at least I hope so.
It’s time to start implementing support for recognizing gestures with neural networks. As I mentioned in the previous post, I had seen some potential problems. After two days of work, I can finally write that those problems are solved. In this post, I’ll describe how I solved the problem of different drawing area location. In the second paragraph, I’m going to describe implemented resizing strategy. At the end, in the third part, I’ll write some words about flattering gestures and values normalization.