I've had some confusion on this for some time now. When FaceNet is run on an image, it returns an 128 element array in Euclidean space/L2 (even this is something I do not completely get). I've had the thought that maybe this embedding can be used to predict faces with triplet loss.
My question is: can I compute something like triplet loss using this embedding? And if so, how is it done? Do I subtract an element corresponding to the other like:
arr = [] #supposed new embedding after difference is calculated
for i in range(0,128):
j = b[i] - a[i]
arr.append[j]
Is this how it's done, and can I perform facial identification with this?
Please kindly move this to the right forum if it's not appropriate here.