VisualActionKit
VisualActionKit copied to clipboard
Human action classification for video, offline and natively on iOS via Core ML
VisualActionKit
Human action classification for video, offline and natively on iOS via Core ML
Uses the kinetics-i3d model to classify videos into one of 400 different action classes defined in Kinetics 400
⚠️ This project requires Xcode 12.
Reference: See accompanying blog post
Installation
To install via Swift Package Manager, add VisualActionKit to your Package.swift file. Alternatively, add it from Xcode directly.
let package = Package(
...
dependencies: [
.package(url: "https://github.com/lukereichold/VisualActionKit.git", from: "0.1.0")
],
...
)
Then import VisualActionKit wherever you’d like to use it:
import VisualActionKit
Usage
let url = Bundle.module.url(forResource: "writing", withExtension: "mp4")
let asset = AVAsset(url: url)
try Classifier.shared.classify(asset) { predictions in
print(predictions)
}
Contribute
Contributions welcome. Please check out the issues.