Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement rules for combining saliency detection with face tracking #151

Open
amebel opened this issue Jan 11, 2017 · 3 comments
Open

Implement rules for combining saliency detection with face tracking #151

amebel opened this issue Jan 11, 2017 · 3 comments
Assignees

Comments

@amebel
Copy link
Contributor

amebel commented Jan 11, 2017

  • We want to distinguish “salient face events” from “salient non-face events" so, then, if there are no faces present, the behavior can be just to do saliency tracking and look around at salient events.

  • At the moment the system cannot recognize anything but faces — when we integrate object recognition software, then maybe it should have some tendency to look at objects it can recognize, as well as the tendency to look at salient stuff. But if there are faces present, then in general the face-tracking behavior should overrule the saliency tracking, unless something REALLY salient pops up.

  • When we make the statement “Han, look at this!” (or in general "ROBOTNAME, look at this!”) cause it to temporarily prioritize looking at salient events nearby the robot. And, “Look at that!” cause it to temporarily prioritize looking at salient events anywhere.

@amebel amebel self-assigned this Jan 11, 2017
@linas
Copy link
Member

linas commented Jan 11, 2017

The code for "look at this" is mostly(?) done. I briefly had some version of it working.

@linas
Copy link
Member

linas commented Jan 11, 2017

It's in the same file as "look at me" and "look at him", and the place where its unfinished is the part that tried to ground "me", "him" and "this" into the same kind of data structure. For some of these, the data structure was a face ID, in other cases, the data structure was a 3D point. So when I said "mostly" working", it works if you pick one or the other data structure, but since half the pipeline uses 3D points, and the other half uses face_id's, both styles could not work together at the same time.

I decided the best solution was to use object-id's everywhere -- i.e. convert face-id into an object-id, and then use object-ids for saliency. However, I only got party-way done in that conversion.

@linas
Copy link
Member

linas commented Jan 11, 2017

Also, "look at him" needed to be integrated with the manyears code -- I think I got part-way through with that, as well, so I could tell apart "me" (where sound is coming from) from "him" (a second visible face that isn't making sound).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants