📜 ⬆️ ⬇️

Dog ate on neural networks

You see you on the street dog. You always see dogs on the street. Someone doesn’t notice them at all and doesn’t have any troubles, these spiritual torments are not there, and in general we can assume that they are on the street of norms. But you! And I noticed, and your eyes are in a wet place, you think like this: “dog. wah, so pretty, I would take it but I can't. But that's right for sure, but nothing. And he went further. But pofotal him, stroked.


With such good little ones, we need to throw geos at the moment and a couple of photos into our system.

There are those who really want to find their runaway to mate, or something else that a baby girl. This one really definitely wants to find a pet. They themselves place ads, walk through doorways, basements and on different sites.

2019 has to more technological solutions. And our project with the working title PetSI (PetSearchInstrument) is exactly that.
')

Concept


In the framework of the Open Data Science community’s Machine Learning for Social Good, we, together with 9851754 and our team, make a service for searching for missing animals, in which the owner can indicate a photo of the animal, the address of the loss and other characteristics, and in response receive the most relevant from our point of view algorithm for reporting found or seen animals.

A brief algorithm for the operation of our service: we aggregate data (photos, location, breed, etc.) from several sites, convert images by a neural network into a vector, train knn, and show the nearest neighbors of the entered photo. You find a lost animal, the animal returns home. Everyone is happy)



In addition to searching for missing pets, we are developing a recommender system to speed up the addition of shelter animals to new owners. While we do not have statistics on user behavior on the site, we therefore use content-based recommendations based on visual similarity.



What already is



Under the hood


Development is conducted in python. We use the following technology stack:


To synchronize all stages of our workflow, we use Airflow:

Multistage data collection. First, the spiders collect information and send it into a queue in its raw form. On the other side of the queue, special handlers convert the data to the desired type (for example, convert the text to an address) and add it to the DBMS.



The collected data is validated and sent to the training model. A special service has been written for the site that loads new data and models, and also initiates recalculation of search results for each user. The whole cycle takes about 8 hours.

How we use machine learning and data analysis:


Immediate plans



Come to us!




For further development we need:


Why it might be necessary for you:


Have a plan. We need people in the team.

Write in a personal or fill in the form and join!

Instead of conclusion


Finally, we prepared a selection of meme dogs and similar dogs from shelters, which we found using our algorithms

Source: https://habr.com/ru/post/459988/


All Articles