OpenALPR integration for React Native. Provides a camera component that recognizes license plates in real-time. Supports both iOS and Android.
- iOS 9+
- Android 5.0+
- RN 0.60+
Start by adding the package and linking it.
$ yarn add react-native-openalpr
$ cd ios && pod install && cd ..
There is workaround for a missing libstdc++.6.0.9.dylib in newer xcode. Please refer to example/ios/Podfile. Remember to set target.name when copy pasting.
There is also an workaround for xcode 12.5 if neeeded.
- Add an entry for
NSCameraUsageDescription
in yourinfo.plist
explaining why your app will use the camera. If you forget to add this, your app will crash!
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
<plist version="1.0">
<dict>
...
<key>NSCameraUsageDescription</key>
<string>We use your camera for license plate recognition to make it easier for you to add your vehicle.</string>
</dict>
Because the OpenCV binary framework release is compiled without bitcode, the other frameworks built by this script are also built without it, which ultimately means your Xcode project also cannot be built with bitcode enabled. Per this message, it sounds like we want this feature disabled for OpenCV anyway.
To disable bitcode in your project:
- In
Build Settings
→Build Options
, search forEnable Bitcode
and set it toNo
.
- Add permissions for
CAMERA
andFLASHLIGHT
and the related features (below) toAndroidManifest.xml
. If you forget to add these permissions, your app will crash!
<manifest xmlns:android="http://schemas.android.com/apk/res/android"
<!-- Camera Permissions -->
<uses-permission android:name="android.permission.CAMERA" />
<uses-feature
android:name="android.hardware.camera"
android:required="false" />
<uses-feature
android:name="android.hardware.camera.autofocus"
android:required="false" />
<uses-permission android:name="android.permission.FLASHLIGHT" />
rootProject.name = 'RNOpenALPRExample'
apply from: file("../node_modules/@react-native-community/cli-platform-android/native_modules.gradle"); applyNativeModulesSettingsGradle(settings)
include ':app'
// Add these lines
include ':openalpr'
project(':openalpr').projectDir = new File(rootProject.projectDir, '../node_modules/react-native-openalpr/android/libraries/openalpr')
include ':opencv'
project(':opencv').projectDir = new File(rootProject.projectDir, '../node_modules/react-native-openalpr/android/libraries/opencv')
The library is linked automatically with leptonica, opencv, tesseract, and openalpr (openalpr).
To make it work, copy and paste the directory with the runtime needed data to your project at path android/app/src/main/assets/runtime_data
.
The runtime_data
file can be found in /example/android/app/src/main/assets/
in this repo. Open runtime_data/openalpr.conf
file and replace com.rnopenalprexample
with your package name
[common]
; Specify the path to the runtime data directory
runtime_dir = /data/data/com.rnopenalprexample/runtime_data
ocr_img_size_percent = 1.33333333
state_id_img_size_percent = 2.0
...
OpenALPR exposes a camera component (based on react-native-camera) that is optimized to run OpenALPR image processing on a live camera stream. Among other parameters, the camera accepts a callback, onPlateRecognized
, for when a plate is recognized.
import React, { Component } from 'react'
import { StyleSheet, Text, View } from 'react-native'
import Camera, {
Aspect,
CaptureQuality,
TorchMode,
} from 'react-native-openalpr'
const styles = StyleSheet.create({
container: {
flex: 1,
},
textContainer: {
position: 'absolute',
top: 100,
left: 50,
},
text: {
textAlign: 'center',
fontSize: 20,
},
})
export default class PlateRecognizer extends React.Component {
state = {
plate: 'Scan a plate',
}
onPlateRecognized = ({ plate, confidence }) => {
this.setState({
plate,
})
}
render() {
return (
<View style={styles.container}>
<Camera
style={styles.preview}
aspect={Aspect.fill}
captureQuality={CaptureQuality.medium}
country="us"
onPlateRecognized={this.onPlateRecognized}
plateOutlineColor="#ff0000"
showPlateOutline
zoom={0}
torchMode={TorchMode.off}
touchToFocus
/>
<View style={styles.textContainer}>
<Text style={styles.text}>{this.state.plate}</Text>
</View>
</View>
)
}
}
The zoon of the camera (Android only). Can be :
0 to 99
The aspect ratio of the camera. Can be one of:
Aspect.stretch
Aspect.fit
Aspect.fill
The resolution at which video frames are captured and analyzed. For completeness, several options are provided. However, it is strongly recommended that you stick with one of the following for the best frame rates and accuracy:
CaptureQuality.medium
(480x360)CaptureQuality.480p
(640x480)
Specifies which OpenALPR config file to load, corresponding to the country whose plates you wish to recognize. Currently supported values are:
au
br
eu
fr
gb
kr
mx
sg
us
vn2
This callback receives a hash with keys:
plate
, representing the recognized license plate stringconfidence
, OpenALPR's confidence(%) in the result
Hex string specifying the color of the border to draw around the recognized plate. Example: #ff0000
for red.
If true, this draws an outline over the recognized plate
Turns the flashlight on or off. Can be one of:
TorchMode.on
TorchMode.off
TorchMode.auto
If true, this focuses the camera where the user taps
- This project works with iOS and Android. It may have some bugs depending on how the underlying native components are updated
- Clone the repo and enter the
example
directory
git clone https://github.com/RobertSasak/react-native-openalpr.git
cd react-native-openalpr
cd example
-
From the
example
directory, runyarn
-
Copy the
android
folder from/react-native-openalpr/android
to/react-native-openalpr/example/node_modules/react-native-openalpr/
-
Open Android Studio and import the project
react-native-openalpr/example/android
and wait until Android Studio indexes and links. -
Run
npm start
from dir /react-native-openalpr/example/ -
Open the path in your browser
http://localhost:8081/index.android.bundle?platform=android&dev=true&hot=false&minify=false
-
Create file the
/react-native-openalpr/example/android/app/src/main/assets/index.android.bundle
. Copy and paste the data from browser window to the file you just created and save. -
Return to Android Studio and run project on your development device.
Note: If you are getting errors, double check that you have completed all of the steps above. If you are having issues running npm start
on Mac OSX and are using homebrew, this issue might help.
- OpenALPR built from OpenALPR-iOS
- Project scaffold based on react-native-camera