您好,登錄后才能下訂單哦!
本篇內容主要講解“Android開發基于ArcSoft實現人臉識別的方法是什么”,感興趣的朋友不妨來看看。本文介紹的方法操作簡單快捷,實用性強。下面就讓小編來帶大家學習“Android開發基于ArcSoft實現人臉識別的方法是什么”吧!
int activeCode = FaceEngine.activeOnline( ChooseFunctionActivity.this, Param.APP_ID, Param.SDK_KEY);
public static final String APP_ID = "AwY6okHQHxtM92YRYSEqJQwb8cED5huPvYyMhK1w7BSo"; public static final String SDK_KEY = "AF8SaLYtP3ALsmaTR55y9UXaykBZjTtMt5gwCBkUGugh";
private boolean checkSoFile(String[] libraries) { File dir = new File(getApplicationInfo().nativeLibraryDir); File[] files = dir.listFiles(); if (files == null || files.length == 0) { return false; } List<String> libraryNameList = new ArrayList<>(); for (File file : files) { libraryNameList.add(file.getName()); } boolean exists = true; for (String library : libraries) { exists &= libraryNameList.contains(library); } return exists; }
protected boolean CheckPermissions(String[] neededPermissions) { if (neededPermissions == null || neededPermissions.length == 0) { return true; } boolean allGranted = true; for (String neededPermission : neededPermissions) { allGranted &= ContextCompat.checkSelfPermission(this, neededPermission) == PackageManager.PERMISSION_GRANTED; } return allGranted; }
激活引擎代碼如下
public void ActivationDevice(final View view) { if (!libraryExists) { ShowToast(getString(R.string.library_not_found)); return; } if (!CheckPermissions(NEEDED_PERMISSIONS)) { ActivityCompat.requestPermissions(this, NEEDED_PERMISSIONS, ACTION_REQUEST_PERMISSIONS); return; } if (view != null) { view.setClickable(false); } Observable.create( new ObservableOnSubscribe<Integer>() { @Override public void subscribe(ObservableEmitter<Integer> emitter) { RuntimeABI runtimeABI = FaceEngine.getRuntimeABI(); Log.i(TAG, "subscribe: getRuntimeABI() " + runtimeABI); long start = System.currentTimeMillis(); int activeCode = FaceEngine.activeOnline( ChooseFunctionActivity.this, Param.APP_ID, Param.SDK_KEY); Log.i(TAG, "subscribe cost: " + (System.currentTimeMillis() - start)); emitter.onNext(activeCode); } }) .subscribeOn( Schedulers.io()) .observeOn( AndroidSchedulers.mainThread()) .subscribe(new Observer<Integer>() { @Override public void onSubscribe(Disposable d) { } @Override public void onNext(Integer activeCode) { if (activeCode == ErrorInfo.MOK) { ShowToast(getString(R.string.activation_succeeded)); } else if (activeCode == ErrorInfo.MERR_ASF_ALREADY_ACTIVATED) { ShowToast(getString(R.string.already_activated)); } else { ShowToast(getString(R.string.active_failed, activeCode)); } if (view != null) { view.setClickable(true); } ActiveFileInfo activeFileInfo = new ActiveFileInfo(); int res = FaceEngine.getActiveFileInfo(ChooseFunctionActivity.this, activeFileInfo); if (res == ErrorInfo.MOK) { Log.i(TAG, activeFileInfo.toString()); } } @Override public void onError(Throwable e) { ShowToast(e.getMessage()); if (view != null) { view.setClickable(true); } } @Override public void onComplete() { } }); }
FaceServer.getInstance().init(this);
public void onGlobalLayout() { previewView.getViewTreeObserver().removeOnGlobalLayoutListener(this); if (!CheckPermissions(NEEDED_PERMISSIONS)) { ActivityCompat.requestPermissions(this, NEEDED_PERMISSIONS, ACTION_REQUEST_PERMISSIONS); } else { initEngine(); initCamera(); } }
private void initEngine() { ftEngine = new FaceEngine(); ftInitCode = ftEngine.init(this, DetectMode.ASF_DETECT_MODE_VIDEO, ConfigUtil.getFtOrient(this), 16, MAX_DETECT_NUM, FaceEngine.ASF_FACE_DETECT); frEngine = new FaceEngine(); frInitCode = frEngine.init(this, DetectMode.ASF_DETECT_MODE_IMAGE, DetectFaceOrientPriority.ASF_OP_0_ONLY, 16, MAX_DETECT_NUM, FaceEngine.ASF_FACE_RECOGNITION); flEngine = new FaceEngine(); flInitCode = flEngine.init(this, DetectMode.ASF_DETECT_MODE_IMAGE, DetectFaceOrientPriority.ASF_OP_0_ONLY, 16, MAX_DETECT_NUM, FaceEngine.ASF_LIVENESS); Log.i(TAG, "initEngine: init: " + ftInitCode); if (ftInitCode != ErrorInfo.MOK) { String error = getString(R.string.specific_engine_init_failed, "ftEngine", ftInitCode); Log.i(TAG, "initEngine: " + error); ShowToast(error); } if (frInitCode != ErrorInfo.MOK) { String error = getString(R.string.specific_engine_init_failed, "frEngine", frInitCode); Log.i(TAG, "initEngine: " + error); ShowToast(error); } if (flInitCode != ErrorInfo.MOK) { String error = getString(R.string.specific_engine_init_failed, "flEngine", flInitCode); Log.i(TAG, "initEngine: " + error); ShowToast(error); } }
private void initCamera() { DisplayMetrics metrics = new DisplayMetrics(); getWindowManager().getDefaultDisplay().getMetrics(metrics); final FaceListener faceListener = new FaceListener() { @Override public void onFail(Exception e) { Log.e(TAG, "onFail: " + e.getMessage()); } //請求FR的回調 @Override public void onFaceFeatureInfoGet(@Nullable final FaceFeature faceFeature, final Integer requestId, final Integer errorCode) { //FR成功 if (faceFeature != null) { // Log.i(TAG, "onPreview: fr end = " + System.currentTimeMillis() + " trackId = " + requestId); Integer liveness = livenessMap.get(requestId); //不做活體檢測的情況,直接搜索 if (!livenessDetect) { searchFace(faceFeature, requestId); } //活體檢測通過,搜索特征 else if (liveness != null && liveness == LivenessInfo.ALIVE) { searchFace(faceFeature, requestId); } //活體檢測未出結果,或者非活體,延遲執行該函數 else { if (requestFeatureStatusMap.containsKey(requestId)) { Observable.timer(WAIT_LIVENESS_INTERVAL, TimeUnit.MILLISECONDS) .subscribe(new Observer<Long>() { Disposable disposable; @Override public void onSubscribe(Disposable d) { disposable = d; getFeatureDelayedDisposables.add(disposable); } @Override public void onNext(Long aLong) { onFaceFeatureInfoGet(faceFeature, requestId, errorCode); } @Override public void onError(Throwable e) { } @Override public void onComplete() { getFeatureDelayedDisposables.remove(disposable); } }); } } } //特征提取失敗 else { if (increaseAndGetValue(extractErrorRetryMap, requestId) > MAX_RETRY_TIME) { extractErrorRetryMap.put(requestId, 0); String msg; // 傳入的FaceInfo在指定的圖像上無法解析人臉,此處使用的是RGB人臉數據,一般是人臉模糊 if (errorCode != null && errorCode == ErrorInfo.MERR_FSDK_FACEFEATURE_LOW_CONFIDENCE_LEVEL) { msg = getString(R.string.low_confidence_level); } else { msg = "ExtractCode:" + errorCode; } faceHelper.setName(requestId, getString(R.string.recognize_failed_notice, msg)); // 在嘗試最大次數后,特征提取仍然失敗,則認為識別未通過 requestFeatureStatusMap.put(requestId, RequestFeatureStatus.FAILED); retryRecognizeDelayed(requestId); } else { requestFeatureStatusMap.put(requestId, RequestFeatureStatus.TO_RETRY); } } } @Override public void onFaceLivenessInfoGet(@Nullable LivenessInfo livenessInfo, final Integer requestId, Integer errorCode) { if (livenessInfo != null) { int liveness = livenessInfo.getLiveness(); livenessMap.put(requestId, liveness); // 非活體,重試 if (liveness == LivenessInfo.NOT_ALIVE) { faceHelper.setName(requestId, getString(R.string.recognize_failed_notice, "NOT_ALIVE")); // 延遲 FAIL_RETRY_INTERVAL 后,將該人臉狀態置為UNKNOWN,幀回調處理時會重新進行活體檢測 retryLivenessDetectDelayed(requestId); } } else { if (increaseAndGetValue(livenessErrorRetryMap, requestId) > MAX_RETRY_TIME) { livenessErrorRetryMap.put(requestId, 0); String msg; // 傳入的FaceInfo在指定的圖像上無法解析人臉,此處使用的是RGB人臉數據,一般是人臉模糊 if (errorCode != null && errorCode == ErrorInfo.MERR_FSDK_FACEFEATURE_LOW_CONFIDENCE_LEVEL) { msg = getString(R.string.low_confidence_level); } else { msg = "ProcessCode:" + errorCode; } faceHelper.setName(requestId, getString(R.string.recognize_failed_notice, msg)); retryLivenessDetectDelayed(requestId); } else { livenessMap.put(requestId, LivenessInfo.UNKNOWN); } } } }; CameraListener cameraListener = new CameraListener() { @Override public void onCameraOpened(Camera camera, int cameraId, int displayOrientation, boolean isMirror) { Camera.Size lastPreviewSize = previewSize; previewSize = camera.getParameters().getPreviewSize(); drawHelper = new DrawHelper(previewSize.width, previewSize.height, previewView.getWidth(), previewView.getHeight(), displayOrientation , cameraId, isMirror, false, false); Log.i(TAG, "onCameraOpened: " + drawHelper.toString()); // 切換相機的時候可能會導致預覽尺寸發生變化 if (faceHelper == null || lastPreviewSize == null || lastPreviewSize.width != previewSize.width || lastPreviewSize.height != previewSize.height) { Integer trackedFaceCount = null; // 記錄切換時的人臉序號 if (faceHelper != null) { trackedFaceCount = faceHelper.getTrackedFaceCount(); faceHelper.release(); } faceHelper = new FaceHelper.Builder() .ftEngine(ftEngine) .frEngine(frEngine) .flEngine(flEngine) .frQueueSize(MAX_DETECT_NUM) .flQueueSize(MAX_DETECT_NUM) .previewSize(previewSize) .faceListener(faceListener) .trackedFaceCount(trackedFaceCount == null ? ConfigUtil.getTrackedFaceCount(FaceComparison_RGB.this.getApplicationContext()) : trackedFaceCount) .build(); } } @Override public void onPreview(final byte[] nv21, Camera camera) { if (faceRectView != null) { faceRectView.clearFaceInfo(); } List<FacePreviewInfo> facePreviewInfoList = faceHelper.onPreviewFrame(nv21); if (facePreviewInfoList != null && faceRectView != null && drawHelper != null) { drawPreviewInfo(facePreviewInfoList); } registerFace(nv21, facePreviewInfoList); clearLeftFace(facePreviewInfoList); if (facePreviewInfoList != null && facePreviewInfoList.size() > 0 && previewSize != null) { for (int i = 0; i < facePreviewInfoList.size(); i++) { Integer status = requestFeatureStatusMap.get(facePreviewInfoList.get(i).getTrackId()); /** * 在活體檢測開啟,在人臉識別狀態不為成功或人臉活體狀態不為處理中(ANALYZING)且不為處理完成(ALIVE、NOT_ALIVE)時重新進行活體檢測 */ if (livenessDetect && (status == null || status != RequestFeatureStatus.SUCCEED)) { Integer liveness = livenessMap.get(facePreviewInfoList.get(i).getTrackId()); if (liveness == null || (liveness != LivenessInfo.ALIVE && liveness != LivenessInfo.NOT_ALIVE && liveness != RequestLivenessStatus.ANALYZING)) { livenessMap.put(facePreviewInfoList.get(i).getTrackId(), RequestLivenessStatus.ANALYZING); faceHelper.requestFaceLiveness(nv21, facePreviewInfoList.get(i).getFaceInfo(), previewSize.width, previewSize.height, FaceEngine.CP_PAF_NV21, facePreviewInfoList.get(i).getTrackId(), LivenessType.RGB); } } /** * 對于每個人臉,若狀態為空或者為失敗,則請求特征提取(可根據需要添加其他判斷以限制特征提取次數), * 特征提取回傳的人臉特征結果在{@link FaceListener#onFaceFeatureInfoGet(FaceFeature, Integer, Integer)}中回傳 */ if (status == null || status == RequestFeatureStatus.TO_RETRY) { requestFeatureStatusMap.put(facePreviewInfoList.get(i).getTrackId(), RequestFeatureStatus.SEARCHING); faceHelper.requestFaceFeature(nv21, facePreviewInfoList.get(i).getFaceInfo(), previewSize.width, previewSize.height, FaceEngine.CP_PAF_NV21, facePreviewInfoList.get(i).getTrackId()); // Log.i(TAG, "onPreview: fr start = " + System.currentTimeMillis() + " trackId = " + facePreviewInfoList.get(i).getTrackedFaceCount()); } } } } @Override public void onCameraClosed() { Log.i(TAG, "onCameraClosed: "); } @Override public void onCameraError(Exception e) { Log.i(TAG, "onCameraError: " + e.getMessage()); } @Override public void onCameraConfigurationChanged(int cameraID, int displayOrientation) { if (drawHelper != null) { drawHelper.setCameraDisplayOrientation(displayOrientation); } Log.i(TAG, "onCameraConfigurationChanged: " + cameraID + " " + displayOrientation); } }; cameraHelper = new CameraHelper.Builder() .previewViewSize(new Point(previewView.getMeasuredWidth(), previewView.getMeasuredHeight())) .rotation(getWindowManager().getDefaultDisplay().getRotation()) .specificCameraId(rgbCameraID != null ? rgbCameraID : Camera.CameraInfo.CAMERA_FACING_FRONT) .isMirror(false) .previewOn(previewView) .cameraListener(cameraListener) .build(); cameraHelper.init(); cameraHelper.start(); }
private void registerFace(final byte[] nv21, final List<FacePreviewInfo> facePreviewInfoList) { if (registerStatus == REGISTER_STATUS_READY && facePreviewInfoList != null && facePreviewInfoList.size() > 0) { registerStatus = REGISTER_STATUS_PROCESSING; Observable.create( new ObservableOnSubscribe<Boolean>() { @Override public void subscribe(ObservableEmitter<Boolean> emitter) { boolean success = FaceServer.getInstance().registerNv21(FaceComparison_RGB.this, nv21.clone(), previewSize.width, previewSize.height, facePreviewInfoList.get(0).getFaceInfo(), "registered" + faceHelper.getTrackedFaceCount()); emitter.onNext(success); } }) .subscribeOn( Schedulers.computation()) .observeOn( AndroidSchedulers.mainThread()) .subscribe(new Observer<Boolean>() { @Override public void onSubscribe(Disposable d) { } /**判斷是否注冊成功*/ @Override public void onNext(Boolean success) { //String result = success ? "register success!" : "register failed!"; //ShowToast(result); // AlertDialog.Builder builder = new AlertDialog.Builder( FaceComparison_RGB.this ); // AlertDialog dialog = builder.create(); // View AlertDialog_View = View.inflate( FaceComparison_RGB.this,R.layout.register_result,null ); // dialog.setView( AlertDialog_View ); // dialog.show(); ShowPopWindows(success); registerStatus = REGISTER_STATUS_DONE; } @Override public void onError(Throwable e) { e.printStackTrace(); ShowToast("register failed!"); ShowFailPopWindows(); registerStatus = REGISTER_STATUS_DONE; } @Override public void onComplete() { } }); } }
public void switchCamera(View view) { if (cameraHelper != null) { boolean success = cameraHelper.switchCamera(); if (!success) { ShowToast(getString(R.string.switch_camera_failed)); } else { ShowToast(getString(R.string.notice_change_detect_degree)); } } }
到此,相信大家對“Android開發基于ArcSoft實現人臉識別的方法是什么”有了更深的了解,不妨來實際操作一番吧!這里是億速云網站,更多相關內容可以進入相關頻道進行查詢,關注我們,繼續學習!
免責聲明:本站發布的內容(圖片、視頻和文字)以原創、轉載和分享為主,文章觀點不代表本網站立場,如果涉及侵權請聯系站長郵箱:is@yisu.com進行舉報,并提供相關證據,一經查實,將立刻刪除涉嫌侵權內容。