mirror of
https://github.com/serengil/deepface.git
synced 2025-06-08 12:35:22 +00:00
ensemble verify becomes cleaner
This commit is contained in:
parent
45f17417af
commit
9f436f39a7
@ -50,33 +50,29 @@ def verify(img1_path, img2_path = '', model_name = 'VGG-Face', distance_metric =
|
|||||||
#--------------------------------
|
#--------------------------------
|
||||||
|
|
||||||
if model_name == 'Ensemble':
|
if model_name == 'Ensemble':
|
||||||
return Boosting.verify(model = model, img_list = img_list, bulkProcess = bulkProcess, enforce_detection = enforce_detection, detector_backend = detector_backend)
|
model_names = ["VGG-Face", "Facenet", "OpenFace", "DeepFace"]
|
||||||
|
metrics = ["cosine", "euclidean", "euclidean_l2"]
|
||||||
#ensemble learning block end
|
else:
|
||||||
|
model_names = []; metrics = []
|
||||||
|
model_names.append(model_name)
|
||||||
|
metrics.append(distance_metric)
|
||||||
|
|
||||||
#--------------------------------
|
#--------------------------------
|
||||||
#ensemble learning disabled
|
#ensemble learning disabled
|
||||||
|
|
||||||
if model == None:
|
if model == None:
|
||||||
model = build_model(model_name)
|
if model_name == 'Ensemble':
|
||||||
|
models = Boosting.loadModel()
|
||||||
#------------------------------
|
|
||||||
#face recognition models have different size of inputs
|
|
||||||
#my environment returns (None, 224, 224, 3) but some people mentioned that they got [(None, 224, 224, 3)]. I think this is because of version issue.
|
|
||||||
|
|
||||||
input_shape = model.layers[0].input_shape
|
|
||||||
|
|
||||||
if type(input_shape) == list:
|
|
||||||
input_shape = input_shape[0][1:3]
|
|
||||||
else:
|
else:
|
||||||
input_shape = input_shape[1:3]
|
model = build_model(model_name)
|
||||||
|
models = {}
|
||||||
input_shape_x = input_shape[0]; input_shape_y = input_shape[1]
|
models[model_name] = model
|
||||||
|
else:
|
||||||
#------------------------------
|
if model_name == 'Ensemble':
|
||||||
|
Boosting.validate_model(model)
|
||||||
#tuned thresholds for model and metric pair
|
else:
|
||||||
threshold = functions.findThreshold(model_name, distance_metric)
|
models = {}
|
||||||
|
models[model_name] = model
|
||||||
|
|
||||||
#------------------------------
|
#------------------------------
|
||||||
|
|
||||||
@ -85,35 +81,51 @@ def verify(img1_path, img2_path = '', model_name = 'VGG-Face', distance_metric =
|
|||||||
|
|
||||||
pbar = tqdm(range(0,len(img_list)), desc='Verification', disable = disable_option)
|
pbar = tqdm(range(0,len(img_list)), desc='Verification', disable = disable_option)
|
||||||
|
|
||||||
#for instance in img_list:
|
|
||||||
for index in pbar:
|
for index in pbar:
|
||||||
|
|
||||||
instance = img_list[index]
|
instance = img_list[index]
|
||||||
|
|
||||||
if type(instance) == list and len(instance) >= 2:
|
if type(instance) == list and len(instance) >= 2:
|
||||||
img1_path = instance[0]
|
img1_path = instance[0]; img2_path = instance[1]
|
||||||
img2_path = instance[1]
|
|
||||||
|
ensemble_features = []
|
||||||
|
|
||||||
|
for i in model_names:
|
||||||
|
custom_model = models[i]
|
||||||
|
|
||||||
|
#decide input shape
|
||||||
|
input_shape = functions.find_input_shape(custom_model)
|
||||||
|
input_shape_x = input_shape[0]; input_shape_y = input_shape[1]
|
||||||
|
|
||||||
#----------------------
|
#----------------------
|
||||||
#crop and align faces
|
#detect and align faces
|
||||||
|
|
||||||
img1 = functions.preprocess_face(img=img1_path, target_size=(input_shape_y, input_shape_x), enforce_detection = enforce_detection, detector_backend = detector_backend)
|
img1 = functions.preprocess_face(img=img1_path
|
||||||
img2 = functions.preprocess_face(img=img2_path, target_size=(input_shape_y, input_shape_x), enforce_detection = enforce_detection, detector_backend = detector_backend)
|
, target_size=(input_shape_y, input_shape_x)
|
||||||
|
, enforce_detection = enforce_detection
|
||||||
|
, detector_backend = detector_backend)
|
||||||
|
|
||||||
|
img2 = functions.preprocess_face(img=img2_path
|
||||||
|
, target_size=(input_shape_y, input_shape_x)
|
||||||
|
, enforce_detection = enforce_detection
|
||||||
|
, detector_backend = detector_backend)
|
||||||
|
|
||||||
#----------------------
|
#----------------------
|
||||||
#find embeddings
|
#find embeddings
|
||||||
|
|
||||||
img1_representation = model.predict(img1)[0,:]
|
img1_representation = custom_model.predict(img1)[0,:]
|
||||||
img2_representation = model.predict(img2)[0,:]
|
img2_representation = custom_model.predict(img2)[0,:]
|
||||||
|
|
||||||
#----------------------
|
#----------------------
|
||||||
#find distances between embeddings
|
#find distances between embeddings
|
||||||
|
|
||||||
if distance_metric == 'cosine':
|
for j in metrics:
|
||||||
|
|
||||||
|
if j == 'cosine':
|
||||||
distance = dst.findCosineDistance(img1_representation, img2_representation)
|
distance = dst.findCosineDistance(img1_representation, img2_representation)
|
||||||
elif distance_metric == 'euclidean':
|
elif j == 'euclidean':
|
||||||
distance = dst.findEuclideanDistance(img1_representation, img2_representation)
|
distance = dst.findEuclideanDistance(img1_representation, img2_representation)
|
||||||
elif distance_metric == 'euclidean_l2':
|
elif j == 'euclidean_l2':
|
||||||
distance = dst.findEuclideanDistance(dst.l2_normalize(img1_representation), dst.l2_normalize(img2_representation))
|
distance = dst.findEuclideanDistance(dst.l2_normalize(img1_representation), dst.l2_normalize(img2_representation))
|
||||||
else:
|
else:
|
||||||
raise ValueError("Invalid distance_metric passed - ", distance_metric)
|
raise ValueError("Invalid distance_metric passed - ", distance_metric)
|
||||||
@ -121,14 +133,15 @@ def verify(img1_path, img2_path = '', model_name = 'VGG-Face', distance_metric =
|
|||||||
#----------------------
|
#----------------------
|
||||||
#decision
|
#decision
|
||||||
|
|
||||||
|
if model_name != 'Ensemble':
|
||||||
|
|
||||||
|
threshold = functions.findThreshold(i, j)
|
||||||
|
|
||||||
if distance <= threshold:
|
if distance <= threshold:
|
||||||
identified = True
|
identified = True
|
||||||
else:
|
else:
|
||||||
identified = False
|
identified = False
|
||||||
|
|
||||||
#----------------------
|
|
||||||
#response object
|
|
||||||
|
|
||||||
resp_obj = {
|
resp_obj = {
|
||||||
"verified": identified
|
"verified": identified
|
||||||
, "distance": distance
|
, "distance": distance
|
||||||
@ -142,6 +155,39 @@ def verify(img1_path, img2_path = '', model_name = 'VGG-Face', distance_metric =
|
|||||||
resp_objects.append(resp_obj)
|
resp_objects.append(resp_obj)
|
||||||
else:
|
else:
|
||||||
return resp_obj
|
return resp_obj
|
||||||
|
|
||||||
|
else: #Ensemble
|
||||||
|
|
||||||
|
#this returns same with OpenFace - euclidean_l2
|
||||||
|
if i == 'OpenFace' and j == 'euclidean':
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
ensemble_features.append(distance)
|
||||||
|
|
||||||
|
#----------------------
|
||||||
|
|
||||||
|
if model_name == 'Ensemble':
|
||||||
|
|
||||||
|
boosted_tree = Boosting.build_gbm()
|
||||||
|
|
||||||
|
prediction = boosted_tree.predict(np.expand_dims(np.array(ensemble_features), axis=0))[0]
|
||||||
|
|
||||||
|
verified = np.argmax(prediction) == 1
|
||||||
|
score = prediction[np.argmax(prediction)]
|
||||||
|
|
||||||
|
resp_obj = {
|
||||||
|
"verified": verified
|
||||||
|
, "score": score
|
||||||
|
, "distance": ensemble_features
|
||||||
|
, "model": ["VGG-Face", "Facenet", "OpenFace", "DeepFace"]
|
||||||
|
, "similarity_metric": ["cosine", "euclidean", "euclidean_l2"]
|
||||||
|
}
|
||||||
|
|
||||||
|
if bulkProcess == True:
|
||||||
|
resp_objects.append(resp_obj)
|
||||||
|
else:
|
||||||
|
return resp_obj
|
||||||
|
|
||||||
#----------------------
|
#----------------------
|
||||||
|
|
||||||
else:
|
else:
|
||||||
@ -313,16 +359,8 @@ def find(img_path, db_path, model_name ='VGG-Face', distance_metric = 'cosine',
|
|||||||
|
|
||||||
#-------------------------------
|
#-------------------------------
|
||||||
|
|
||||||
#model metric pairs for ensemble
|
|
||||||
model_names = ['VGG-Face', 'Facenet', 'OpenFace', 'DeepFace']
|
|
||||||
metric_names = ['cosine', 'euclidean', 'euclidean_l2']
|
|
||||||
|
|
||||||
#-------------------------------
|
|
||||||
|
|
||||||
if os.path.isdir(db_path) == True:
|
if os.path.isdir(db_path) == True:
|
||||||
|
|
||||||
#---------------------------------------
|
|
||||||
|
|
||||||
if model == None:
|
if model == None:
|
||||||
|
|
||||||
if model_name == 'Ensemble':
|
if model_name == 'Ensemble':
|
||||||
@ -331,14 +369,28 @@ def find(img_path, db_path, model_name ='VGG-Face', distance_metric = 'cosine',
|
|||||||
|
|
||||||
else: #model is not ensemble
|
else: #model is not ensemble
|
||||||
model = build_model(model_name)
|
model = build_model(model_name)
|
||||||
|
models = {}
|
||||||
|
models[model_name] = model
|
||||||
|
|
||||||
else: #model != None
|
else: #model != None
|
||||||
print("Already built model is passed")
|
print("Already built model is passed")
|
||||||
|
|
||||||
if model_name == 'Ensemble':
|
if model_name == 'Ensemble':
|
||||||
|
|
||||||
Boosting.validate_model(model)
|
Boosting.validate_model(model)
|
||||||
models = model.copy()
|
models = model.copy()
|
||||||
|
else:
|
||||||
|
models = {}
|
||||||
|
models[model_name] = model
|
||||||
|
|
||||||
|
#---------------------------------------
|
||||||
|
|
||||||
|
if model_name == 'Ensemble':
|
||||||
|
model_names = ['VGG-Face', 'Facenet', 'OpenFace', 'DeepFace']
|
||||||
|
metric_names = ['cosine', 'euclidean', 'euclidean_l2']
|
||||||
|
elif model_name != 'Ensemble':
|
||||||
|
model_names = []; metric_names = []
|
||||||
|
model_names.append(model_name)
|
||||||
|
metric_names.append(distance_metric)
|
||||||
|
|
||||||
#---------------------------------------
|
#---------------------------------------
|
||||||
|
|
||||||
@ -354,7 +406,7 @@ def find(img_path, db_path, model_name ='VGG-Face', distance_metric = 'cosine',
|
|||||||
|
|
||||||
print("There are ", len(representations)," representations found in ",file_name)
|
print("There are ", len(representations)," representations found in ",file_name)
|
||||||
|
|
||||||
else:
|
else: #create representation.pkl from scratch
|
||||||
employees = []
|
employees = []
|
||||||
|
|
||||||
for r, d, f in os.walk(db_path): # r=root, d=directories, f = files
|
for r, d, f in os.walk(db_path): # r=root, d=directories, f = files
|
||||||
@ -377,47 +429,26 @@ def find(img_path, db_path, model_name ='VGG-Face', distance_metric = 'cosine',
|
|||||||
for index in pbar:
|
for index in pbar:
|
||||||
employee = employees[index]
|
employee = employees[index]
|
||||||
|
|
||||||
if model_name != 'Ensemble':
|
|
||||||
|
|
||||||
#input_shape = model.layers[0].input_shape[1:3] #my environment returns (None, 224, 224, 3) but some people mentioned that they got [(None, 224, 224, 3)]. I think this is because of version issue.
|
|
||||||
|
|
||||||
input_shape = model.layers[0].input_shape
|
|
||||||
|
|
||||||
if type(input_shape) == list:
|
|
||||||
input_shape = input_shape[0][1:3]
|
|
||||||
else:
|
|
||||||
input_shape = input_shape[1:3]
|
|
||||||
|
|
||||||
input_shape_x = input_shape[0]; input_shape_y = input_shape[1]
|
|
||||||
|
|
||||||
img = functions.preprocess_face(img = employee, target_size = (input_shape_y, input_shape_x), enforce_detection = enforce_detection, detector_backend = detector_backend)
|
|
||||||
representation = model.predict(img)[0,:]
|
|
||||||
|
|
||||||
instance = []
|
|
||||||
instance.append(employee)
|
|
||||||
instance.append(representation)
|
|
||||||
|
|
||||||
else: #ensemble learning
|
|
||||||
|
|
||||||
instance = []
|
instance = []
|
||||||
instance.append(employee)
|
instance.append(employee)
|
||||||
|
|
||||||
for j in model_names:
|
for j in model_names:
|
||||||
ensemble_model = models[j]
|
custom_model = models[j]
|
||||||
|
|
||||||
#input_shape = model.layers[0].input_shape[1:3] #my environment returns (None, 224, 224, 3) but some people mentioned that they got [(None, 224, 224, 3)]. I think this is because of version issue.
|
#----------------------------------
|
||||||
|
#decide input shape
|
||||||
input_shape = ensemble_model.layers[0].input_shape
|
|
||||||
|
|
||||||
if type(input_shape) == list:
|
|
||||||
input_shape = input_shape[0][1:3]
|
|
||||||
else:
|
|
||||||
input_shape = input_shape[1:3]
|
|
||||||
|
|
||||||
|
input_shape = functions.find_input_shape(custom_model)
|
||||||
input_shape_x = input_shape[0]; input_shape_y = input_shape[1]
|
input_shape_x = input_shape[0]; input_shape_y = input_shape[1]
|
||||||
|
|
||||||
img = functions.preprocess_face(img = employee, target_size = (input_shape_y, input_shape_x), enforce_detection = enforce_detection, detector_backend = detector_backend)
|
#----------------------------------
|
||||||
representation = ensemble_model.predict(img)[0,:]
|
|
||||||
|
img = functions.preprocess_face(img = employee
|
||||||
|
, target_size = (input_shape_y, input_shape_x)
|
||||||
|
, enforce_detection = enforce_detection
|
||||||
|
, detector_backend = detector_backend)
|
||||||
|
|
||||||
|
representation = custom_model.predict(img)[0,:]
|
||||||
instance.append(representation)
|
instance.append(representation)
|
||||||
|
|
||||||
#-------------------------------
|
#-------------------------------
|
||||||
@ -431,14 +462,18 @@ def find(img_path, db_path, model_name ='VGG-Face', distance_metric = 'cosine',
|
|||||||
print("Representations stored in ",db_path,"/",file_name," file. Please delete this file when you add new identities in your database.")
|
print("Representations stored in ",db_path,"/",file_name," file. Please delete this file when you add new identities in your database.")
|
||||||
|
|
||||||
#----------------------------
|
#----------------------------
|
||||||
#we got representations for database
|
#now, we got representations for facial database
|
||||||
|
|
||||||
if model_name != 'Ensemble':
|
if model_name != 'Ensemble':
|
||||||
df = pd.DataFrame(representations, columns = ["identity", "representation"])
|
df = pd.DataFrame(representations, columns = ["identity", "%s_representation" % (model_name)])
|
||||||
else: #ensemble learning
|
else: #ensemble learning
|
||||||
df = pd.DataFrame(representations, columns = ["identity", "VGG-Face_representation", "Facenet_representation", "OpenFace_representation", "DeepFace_representation"])
|
|
||||||
|
|
||||||
df_base = df.copy()
|
columns = ['identity']
|
||||||
|
[columns.append('%s_representation' % i) for i in model_names]
|
||||||
|
|
||||||
|
df = pd.DataFrame(representations, columns = columns)
|
||||||
|
|
||||||
|
df_base = df.copy() #df will be filtered in each img. we will restore it for the next item.
|
||||||
|
|
||||||
resp_obj = []
|
resp_obj = []
|
||||||
|
|
||||||
@ -448,21 +483,20 @@ def find(img_path, db_path, model_name ='VGG-Face', distance_metric = 'cosine',
|
|||||||
|
|
||||||
#find representation for passed image
|
#find representation for passed image
|
||||||
|
|
||||||
if model_name == 'Ensemble':
|
|
||||||
for j in model_names:
|
for j in model_names:
|
||||||
ensemble_model = models[j]
|
custom_model = models[j]
|
||||||
|
|
||||||
#input_shape = ensemble_model.layers[0].input_shape[1:3] #my environment returns (None, 224, 224, 3) but some people mentioned that they got [(None, 224, 224, 3)]. I think this is because of version issue.
|
#--------------------------------
|
||||||
|
#decide input shape
|
||||||
|
input_shape = functions.find_input_shape(custom_model)
|
||||||
|
|
||||||
input_shape = ensemble_model.layers[0].input_shape
|
#--------------------------------
|
||||||
|
|
||||||
if type(input_shape) == list:
|
img = functions.preprocess_face(img = img_path, target_size = input_shape
|
||||||
input_shape = input_shape[0][1:3]
|
, enforce_detection = enforce_detection
|
||||||
else:
|
, detector_backend = detector_backend)
|
||||||
input_shape = input_shape[1:3]
|
|
||||||
|
|
||||||
img = functions.preprocess_face(img = img_path, target_size = input_shape, enforce_detection = enforce_detection, detector_backend = detector_backend)
|
target_representation = custom_model.predict(img)[0,:]
|
||||||
target_representation = ensemble_model.predict(img)[0,:]
|
|
||||||
|
|
||||||
for k in metric_names:
|
for k in metric_names:
|
||||||
distances = []
|
distances = []
|
||||||
@ -478,31 +512,45 @@ def find(img_path, db_path, model_name ='VGG-Face', distance_metric = 'cosine',
|
|||||||
|
|
||||||
distances.append(distance)
|
distances.append(distance)
|
||||||
|
|
||||||
if j == 'OpenFace' and k == 'euclidean':
|
#---------------------------
|
||||||
|
|
||||||
|
if model_name == 'Ensemble' and j == 'OpenFace' and k == 'euclidean':
|
||||||
continue
|
continue
|
||||||
else:
|
else:
|
||||||
df["%s_%s" % (j, k)] = distances
|
df["%s_%s" % (j, k)] = distances
|
||||||
|
|
||||||
|
if model_name != 'Ensemble':
|
||||||
|
threshold = functions.findThreshold(j, k)
|
||||||
|
df = df.drop(columns = ["%s_representation" % (j)])
|
||||||
|
df = df[df["%s_%s" % (j, k)] <= threshold]
|
||||||
|
|
||||||
|
df = df.sort_values(by = ["%s_%s" % (j, k)], ascending=True).reset_index(drop=True)
|
||||||
|
|
||||||
|
resp_obj.append(df)
|
||||||
|
df = df_base.copy() #restore df for the next iteration
|
||||||
|
|
||||||
#----------------------------------
|
#----------------------------------
|
||||||
|
|
||||||
|
if model_name == 'Ensemble':
|
||||||
|
|
||||||
feature_names = []
|
feature_names = []
|
||||||
for j in model_names:
|
for j in model_names:
|
||||||
for k in metric_names:
|
for k in metric_names:
|
||||||
if j == 'OpenFace' and k == 'euclidean':
|
if model_name == 'Ensemble' and j == 'OpenFace' and k == 'euclidean':
|
||||||
continue
|
continue
|
||||||
else:
|
else:
|
||||||
feature = '%s_%s' % (j, k)
|
feature = '%s_%s' % (j, k)
|
||||||
feature_names.append(feature)
|
feature_names.append(feature)
|
||||||
|
|
||||||
#print(df[feature_names].head())
|
#print(df.head())
|
||||||
|
|
||||||
x = df[feature_names].values
|
x = df[feature_names].values
|
||||||
|
|
||||||
#----------------------------------
|
#--------------------------------------
|
||||||
#lightgbm model
|
|
||||||
deepface_ensemble = Boosting.build_gbm()
|
|
||||||
|
|
||||||
y = deepface_ensemble.predict(x)
|
boosted_tree = Boosting.build_gbm()
|
||||||
|
|
||||||
|
y = boosted_tree.predict(x)
|
||||||
|
|
||||||
verified_labels = []; scores = []
|
verified_labels = []; scores = []
|
||||||
for i in y:
|
for i in y:
|
||||||
@ -525,49 +573,6 @@ def find(img_path, db_path, model_name ='VGG-Face', distance_metric = 'cosine',
|
|||||||
|
|
||||||
#----------------------------------
|
#----------------------------------
|
||||||
|
|
||||||
if model_name != 'Ensemble':
|
|
||||||
|
|
||||||
#input_shape = model.layers[0].input_shape[1:3] #my environment returns (None, 224, 224, 3) but some people mentioned that they got [(None, 224, 224, 3)]. I think this is because of version issue.
|
|
||||||
|
|
||||||
input_shape = model.layers[0].input_shape
|
|
||||||
|
|
||||||
if type(input_shape) == list:
|
|
||||||
input_shape = input_shape[0][1:3]
|
|
||||||
else:
|
|
||||||
input_shape = input_shape[1:3]
|
|
||||||
|
|
||||||
input_shape_x = input_shape[0]; input_shape_y = input_shape[1]
|
|
||||||
|
|
||||||
#------------------------
|
|
||||||
|
|
||||||
img = functions.preprocess_face(img = img_path, target_size = (input_shape_y, input_shape_x), enforce_detection = enforce_detection, detector_backend = detector_backend)
|
|
||||||
target_representation = model.predict(img)[0,:]
|
|
||||||
|
|
||||||
distances = []
|
|
||||||
for index, instance in df.iterrows():
|
|
||||||
source_representation = instance["representation"]
|
|
||||||
|
|
||||||
if distance_metric == 'cosine':
|
|
||||||
distance = dst.findCosineDistance(source_representation, target_representation)
|
|
||||||
elif distance_metric == 'euclidean':
|
|
||||||
distance = dst.findEuclideanDistance(source_representation, target_representation)
|
|
||||||
elif distance_metric == 'euclidean_l2':
|
|
||||||
distance = dst.findEuclideanDistance(dst.l2_normalize(source_representation), dst.l2_normalize(target_representation))
|
|
||||||
else:
|
|
||||||
raise ValueError("Invalid distance_metric passed - ", distance_metric)
|
|
||||||
|
|
||||||
distances.append(distance)
|
|
||||||
|
|
||||||
threshold = functions.findThreshold(model_name, distance_metric)
|
|
||||||
|
|
||||||
df["distance"] = distances
|
|
||||||
df = df.drop(columns = ["representation"])
|
|
||||||
df = df[df.distance <= threshold]
|
|
||||||
|
|
||||||
df = df.sort_values(by = ["distance"], ascending=True).reset_index(drop=True)
|
|
||||||
resp_obj.append(df)
|
|
||||||
df = df_base.copy() #restore df for the next iteration
|
|
||||||
|
|
||||||
toc = time.time()
|
toc = time.time()
|
||||||
|
|
||||||
print("find function lasts ",toc-tic," seconds")
|
print("find function lasts ",toc-tic," seconds")
|
||||||
|
@ -56,124 +56,3 @@ def build_gbm():
|
|||||||
deepface_ensemble = lgb.Booster(model_file = ensemble_model_path)
|
deepface_ensemble = lgb.Booster(model_file = ensemble_model_path)
|
||||||
|
|
||||||
return deepface_ensemble
|
return deepface_ensemble
|
||||||
|
|
||||||
def verify(model, img_list, bulkProcess, enforce_detection, detector_backend):
|
|
||||||
print("Ensemble learning enabled")
|
|
||||||
|
|
||||||
if model == None:
|
|
||||||
model = loadModel()
|
|
||||||
|
|
||||||
validate_model(model)
|
|
||||||
|
|
||||||
#--------------------------
|
|
||||||
|
|
||||||
model_names = ["VGG-Face", "Facenet", "OpenFace", "DeepFace"]
|
|
||||||
metrics = ["cosine", "euclidean", "euclidean_l2"]
|
|
||||||
|
|
||||||
resp_objects = []
|
|
||||||
|
|
||||||
#--------------------------
|
|
||||||
|
|
||||||
if model == None:
|
|
||||||
model = loadModel()
|
|
||||||
|
|
||||||
#--------------------------
|
|
||||||
|
|
||||||
validate_model(model)
|
|
||||||
|
|
||||||
#--------------------------
|
|
||||||
|
|
||||||
pbar = tqdm(range(0,len(img_list)), desc='Verification')
|
|
||||||
|
|
||||||
for index in pbar:
|
|
||||||
instance = img_list[index]
|
|
||||||
|
|
||||||
if type(instance) == list and len(instance) >= 2:
|
|
||||||
img1_path = instance[0]
|
|
||||||
img2_path = instance[1]
|
|
||||||
|
|
||||||
ensemble_features = []; ensemble_features_string = "["
|
|
||||||
|
|
||||||
for i in model_names:
|
|
||||||
custom_model = model[i]
|
|
||||||
|
|
||||||
input_shape = custom_model.layers[0].input_shape
|
|
||||||
|
|
||||||
if type(input_shape) == list:
|
|
||||||
input_shape = input_shape[0][1:3]
|
|
||||||
else:
|
|
||||||
input_shape = input_shape[1:3]
|
|
||||||
|
|
||||||
#----------------------------------
|
|
||||||
|
|
||||||
img1 = functions.preprocess_face(img = img1_path, target_size = input_shape
|
|
||||||
, enforce_detection = enforce_detection
|
|
||||||
, detector_backend = detector_backend)
|
|
||||||
|
|
||||||
img2 = functions.preprocess_face(img = img2_path, target_size = input_shape
|
|
||||||
, enforce_detection = enforce_detection
|
|
||||||
, detector_backend = detector_backend)
|
|
||||||
|
|
||||||
img1_representation = custom_model.predict(img1)[0,:]
|
|
||||||
img2_representation = custom_model.predict(img2)[0,:]
|
|
||||||
|
|
||||||
for j in metrics:
|
|
||||||
if j == 'cosine':
|
|
||||||
distance = dst.findCosineDistance(img1_representation, img2_representation)
|
|
||||||
elif j == 'euclidean':
|
|
||||||
distance = dst.findEuclideanDistance(img1_representation, img2_representation)
|
|
||||||
elif j == 'euclidean_l2':
|
|
||||||
distance = dst.findEuclideanDistance(dst.l2_normalize(img1_representation), dst.l2_normalize(img2_representation))
|
|
||||||
|
|
||||||
#------------------------
|
|
||||||
|
|
||||||
#this returns same with OpenFace - euclidean_l2
|
|
||||||
if i == 'OpenFace' and j == 'euclidean':
|
|
||||||
continue
|
|
||||||
else:
|
|
||||||
ensemble_features.append(distance)
|
|
||||||
|
|
||||||
if len(ensemble_features) > 1:
|
|
||||||
ensemble_features_string += ", "
|
|
||||||
|
|
||||||
ensemble_features_string += str(distance)
|
|
||||||
|
|
||||||
ensemble_features_string += "]"
|
|
||||||
|
|
||||||
#-------------------------------
|
|
||||||
|
|
||||||
deepface_ensemble = build_gbm()
|
|
||||||
|
|
||||||
#-------------------------------
|
|
||||||
|
|
||||||
prediction = deepface_ensemble.predict(np.expand_dims(np.array(ensemble_features), axis=0))[0]
|
|
||||||
|
|
||||||
verified = np.argmax(prediction) == 1
|
|
||||||
|
|
||||||
score = prediction[np.argmax(prediction)]
|
|
||||||
|
|
||||||
#print("verified: ", verified,", score: ", score)
|
|
||||||
|
|
||||||
resp_obj = {
|
|
||||||
"verified": verified
|
|
||||||
, "score": score
|
|
||||||
, "distance": ensemble_features_string
|
|
||||||
, "model": ["VGG-Face", "Facenet", "OpenFace", "DeepFace"]
|
|
||||||
, "similarity_metric": ["cosine", "euclidean", "euclidean_l2"]
|
|
||||||
}
|
|
||||||
|
|
||||||
if bulkProcess == True:
|
|
||||||
resp_objects.append(resp_obj)
|
|
||||||
else:
|
|
||||||
return resp_obj
|
|
||||||
|
|
||||||
if bulkProcess == True:
|
|
||||||
resp_obj = {}
|
|
||||||
|
|
||||||
for i in range(0, len(resp_objects)):
|
|
||||||
resp_item = resp_objects[i]
|
|
||||||
resp_obj["pair_%d" % (i+1)] = resp_item
|
|
||||||
|
|
||||||
return resp_obj
|
|
||||||
|
|
||||||
|
|
@ -418,6 +418,8 @@ def align_face(img, detector_backend = 'opencv'):
|
|||||||
|
|
||||||
def preprocess_face(img, target_size=(224, 224), grayscale = False, enforce_detection = True, detector_backend = 'opencv'):
|
def preprocess_face(img, target_size=(224, 224), grayscale = False, enforce_detection = True, detector_backend = 'opencv'):
|
||||||
|
|
||||||
|
#img_path = copy.copy(img)
|
||||||
|
|
||||||
#img might be path, base64 or numpy array. Convert it to numpy whatever it is.
|
#img might be path, base64 or numpy array. Convert it to numpy whatever it is.
|
||||||
img = load_image(img)
|
img = load_image(img)
|
||||||
base_img = img.copy()
|
base_img = img.copy()
|
||||||
@ -447,3 +449,18 @@ def preprocess_face(img, target_size=(224, 224), grayscale = False, enforce_dete
|
|||||||
img_pixels /= 255 #normalize input in [0, 1]
|
img_pixels /= 255 #normalize input in [0, 1]
|
||||||
|
|
||||||
return img_pixels
|
return img_pixels
|
||||||
|
|
||||||
|
def find_input_shape(model):
|
||||||
|
|
||||||
|
#face recognition models have different size of inputs
|
||||||
|
#my environment returns (None, 224, 224, 3) but some people mentioned that they got [(None, 224, 224, 3)]. I think this is because of version issue.
|
||||||
|
|
||||||
|
input_shape = model.layers[0].input_shape
|
||||||
|
|
||||||
|
if type(input_shape) == list:
|
||||||
|
input_shape = input_shape[0][1:3]
|
||||||
|
else:
|
||||||
|
input_shape = input_shape[1:3]
|
||||||
|
|
||||||
|
return input_shape
|
||||||
|
|
||||||
|
Binary file not shown.
Before Width: | Height: | Size: 99 KiB |
@ -35,7 +35,7 @@ print(res)
|
|||||||
|
|
||||||
print("-----------------------------------------")
|
print("-----------------------------------------")
|
||||||
|
|
||||||
print("Large scale face recognition")
|
print("Single find function test")
|
||||||
|
|
||||||
df = DeepFace.find(img_path = "dataset/img1.jpg", db_path = "dataset"
|
df = DeepFace.find(img_path = "dataset/img1.jpg", db_path = "dataset"
|
||||||
#, model_name = 'Dlib'
|
#, model_name = 'Dlib'
|
||||||
@ -44,7 +44,30 @@ print(df.head())
|
|||||||
|
|
||||||
print("-----------------------------------------")
|
print("-----------------------------------------")
|
||||||
|
|
||||||
print("Bulk face recognition tests")
|
print("Pre-built model for single find function test")
|
||||||
|
|
||||||
|
model_name = "VGG-Face"
|
||||||
|
model = DeepFace.build_model(model_name)
|
||||||
|
print(model_name," is built")
|
||||||
|
|
||||||
|
df = DeepFace.find(img_path = "dataset/img1.jpg", db_path = "dataset"
|
||||||
|
, model_name = model_name, model = model
|
||||||
|
)
|
||||||
|
print(df.head())
|
||||||
|
|
||||||
|
print("-----------------------------------------")
|
||||||
|
|
||||||
|
print("Bulk find function tests")
|
||||||
|
|
||||||
|
dfs = DeepFace.find(img_path = ["dataset/img1.jpg", "dataset/img2.jpg"], db_path = "dataset"
|
||||||
|
#, model_name = 'Dlib'
|
||||||
|
)
|
||||||
|
print(dfs[0].head())
|
||||||
|
print(dfs[1].head())
|
||||||
|
|
||||||
|
print("-----------------------------------------")
|
||||||
|
|
||||||
|
print("Bulk verification tests")
|
||||||
|
|
||||||
resp_obj = DeepFace.verify(dataset)
|
resp_obj = DeepFace.verify(dataset)
|
||||||
print(resp_obj)
|
print(resp_obj)
|
||||||
@ -241,4 +264,3 @@ print(df)
|
|||||||
|
|
||||||
#-----------------------------------
|
#-----------------------------------
|
||||||
print("--------------------------")
|
print("--------------------------")
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user