一、二维码扫描
1、关键点
(0)框架:AVFoundation
(1)输入设备:摄像头
(2)输出设备:元数据,将二维码解析成字符串输出
(3)建立会话session :通过“添加”,将输入和输出联系起来
(4)预览视图layer : 特殊的layer,专门用来显示输入设备捕捉到的画面
(5)session需要开启和关闭
2、二维码的实质:字符串
3、输出设备output的代理AVCaptureMetadataOutputObjectsDelegate(就一个代理方法)
(1)解析完毕之后就会调用返回一个字符串
- (void)captureOutput:(AVCaptureOutput *)captureOutput didOutputMetadataObjects:(NSArray *)metadataObjects fromConnection:(AVCaptureConnection *)connection;
(2)设置代理的方式很特殊
[self.output setMetadataObjectsDelegate:self queue:dispatch_get_global_queue(DISPATCH_QUEUE_PRIORITY_DEFAULT, 0)];
4、配置示例
#import "ViewController.h" #import <AVFoundation/AVFoundation.h> @interface ViewController ()<AVCaptureMetadataOutputObjectsDelegate> //1.创建输入设备 @property (nonatomic, strong)AVCaptureDeviceInput *deviceInput; //2.创建输出设备 元数据 @property (nonatomic, strong) AVCaptureMetadataOutput *output; //3.创建会话 @property (nonatomic, strong) AVCaptureSession *session; //4.预览视图layer特殊的layer专门来显示输入设备捕捉到的画面 @property (nonatomic, strong) AVCaptureVideoPreviewLayer *preViewLayer; @end @implementation ViewController - (void)viewDidLoad { [super viewDidLoad]; //1.创建输入设备:可以包括摄像头、键盘、鼠标、麦克风 //摄像头有可能有两个 AVCaptureDevice *deviceC = [AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeVideo]; self.deviceInput = [[AVCaptureDeviceInput alloc]initWithDevice:deviceC error:nil]; //2.创建输出设备 元数据 self.output = [[AVCaptureMetadataOutput alloc]init]; // 输出设备设置代理 [self.output setMetadataObjectsDelegate:self queue:dispatch_get_global_queue(DISPATCH_QUEUE_PRIORITY_DEFAULT, 0)]; //3.创建会话 self.session = [[AVCaptureSession alloc]init]; //4.建立联系(判断是否可以添加输入和输出设备) if ([self.session canAddInput:self.deviceInput]) { [self.session addInput:self.deviceInput]; } if ([self.session canAddOutput:self.output]) { [self.session addOutput:self.output]; } //设置元数据类型为“二维码”,还有很多类型,比如条形码 [self.output setMetadataObjectTypes:@[AVMetadataObjectTypeQRCode]]; NSLog(@"%@",self.output.availableMetadataObjectTypes); //视频输出质量 [self.session setSessionPreset:AVCaptureSessionPresetHigh]; //5.开启会话 [self.session startRunning]; //7.增加一个预览视图layer 来展示输入设备的画面 self.preViewLayer = [[AVCaptureVideoPreviewLayer alloc]initWithSession:self.session]; self.preViewLayer.frame = self.view.bounds; //添加到视图 [self.view.layer addSublayer:self.preViewLayer]; } //6.解析完毕之后就会调用:返回一个字符串 - (void)captureOutput:(AVCaptureOutput *)captureOutput didOutputMetadataObjects:(NSArray *)metadataObjects fromConnection:(AVCaptureConnection *)connection; { //取出扫描之后返回的数据 AVMetadataMachineReadableCodeObject *objc = [metadataObjects firstObject]; NSLog(@"%@",objc.stringValue); //停止会话 [self.session stopRunning]; [self.preViewLayer removeFromSuperlayer]; } @end
二、二维码生成
1、使用框架:CoreImage
2、核心工具:滤镜 CIFilter
3、配置示例
#import "ViewController.h" #import <CoreImage/CoreImage.h> @interface ViewController () @property (weak, nonatomic) IBOutlet UIImageView *imageView; @end @implementation ViewController - (void)viewDidLoad { [super viewDidLoad]; //滤镜 //获取内置滤镜支持的种类:二维码 NSLog(@"%@",[CIFilter filterNamesInCategory:kCICategoryBuiltIn]); CIFilter *filter = [CIFilter filterWithName:@"CIQRCodeGenerator"]; //设置默认值 [filter setDefaults]; //封装数据 字符串 NSLog(@"%@",filter.inputKeys); [filter setValue:[@"http://www.baidu.com" dataUsingEncoding:NSUTF8StringEncoding] forKey:@"inputMessage"]; //放大原图片,使得二维码图片更清晰 CIImage *resultImage = [filter.outputImage imageByApplyingTransform:CGAffineTransformMakeScale(5, 5)]; //将滤镜生成的图片赋值给UI控件 self.imageView.image = [UIImage imageWithCIImage:resultImage]; } @end
三、加速计
1、作用:用于检测设备的运动(比如摇晃)
2、加速计原理
(1)检测设备在X、Y、Z轴上的加速度 (哪个方向有力的作用,哪个方向运动了)
(2)根据加速度数值,就可以判断出在各个方向上的作用力度
3、加速计程序的开发
(1)在iOS4以前:使用UIAccelerometer,用法非常简单(到了iOS5就已经过期,UIKit框架)
(2)从iOS4开始:CoreMotion.framework
(3)虽然UIAccelerometer已经过期,但由于其用法极其简单,很多程序里面都还有残留
4、关于CoreMotion
(1)随着iPhone4的推出,加速度计全面升级,并引入了陀螺仪,与Motion(运动)相关的编程成为重头戏,苹果特地在iOS4中增加了专门处理Motion的框架-CoreMotion.framework
(2)Core Motion不仅能够提供实时的加速度值和旋转速度值,更重要的是,苹果在其中集成了很多牛逼的算法
5、UIAccelerometer使用示例(代理方法获得加速计返回的值)
@interface ViewController () <UIAccelerometerDelegate> @property (nonatomic, strong) UIAccelerometer *accelerometer; @end @implementation ViewController - (void)viewDidLoad { [super viewDidLoad]; //检测手机三个方向上的加速度值,注意y的方向向上为正 //重力加速度 //加速计设备是一个硬件设备,单例 UIAccelerometer *accelerometer = [UIAccelerometer sharedAccelerometer]; //设置采样间隔 accelerometer.updateInterval = 1/30.0; //设置代理 accelerometer.delegate = self; self.accelerometer = accelerometer; } //更新了加速度的值就会调用此方法: 参数1 加速计 参数2 返回值 - (void)accelerometer:(UIAccelerometer *)accelerometer didAccelerate:(UIAcceleration *)acceleration { NSLog(@"%f %f %f",acceleration.x,acceleration.y,acceleration.z); } @end
6、CoreMotion框架实现加速计案例:小球滚动
(1)导入框架,创建CMMotionManager
(2)判断加速计是否可用(最好判断):isAccelerometerAvailable
(3)开始采样:startAccelerometerUpdatesToQueue ,采样结果在回调中:accelerometerData
(4)案例源码 (类扩展、工具类、图片、音效文件没有展示)
#import "ViewController.h" #import <CoreMotion/CoreMotion.h> #import "UIView+Extension.h" #import "AudioTool.h" @interface ViewController () @property(nonatomic,strong) CMMotionManager * manager; @property(nonatomic,strong) UIImageView * ball ; @property(nonatomic,assign) CGPoint speed; //记录小球上一刻的位置 @property(nonatomic,assign) CGPoint location; @end @implementation ViewController - (void)viewDidLoad { [super viewDidLoad]; //创建manager CMMotionManager * manager = [[CMMotionManager alloc]init]; self.manager = manager; self.manager.accelerometerUpdateInterval = 1/30.0; //创建小球 self.ball = [[UIImageView alloc]initWithImage:[UIImage imageNamed:@"bb_ball"]]; self.ball.frame = CGRectMake(50, 50, 50, 50); UIImageView * back = [[UIImageView alloc]initWithImage:[UIImage imageNamed:@"bb_background.jpg"]]; back.frame = CGRectMake(0, 0, [UIScreen mainScreen].bounds.size.width, [UIScreen mainScreen].bounds.size.height); [self.view addSubview:back]; [self.view addSubview:self.ball]; //获得加速计数据 [manager startAccelerometerUpdatesToQueue:[NSOperationQueue mainQueue] withHandler:^(CMAccelerometerData * _Nullable accelerometerData, NSError * _Nullable error) { //当前时刻的速度,瞬时速度 = 瞬时距离 _speed.x += accelerometerData.acceleration.x; _speed.y -= accelerometerData.acceleration.y; //更新小球的位置 self.ball.x += _speed.x; self.ball.y += _speed.y; //进行小球的碰撞检测,如果碰到边界,重置小球位置到边界,并且把速度反向,且缩小为一半 if (self.ball.x <= 0) { self.ball.x = 0; _speed.x *= -0.5; } if (self.ball.x >= self.view.width - self.ball.width) { self.ball.x = self.view.width - self.ball.width; _speed.x *= -0.5; } if (self.ball.y <= 0) { self.ball.y = 0; _speed.y *= -0.5; } if (self.ball.y >= self.view.height - self.ball.height) { self.ball.y = self.view.height - self.ball.height; _speed.y *= -0.5; } //碰到边界需要发出声音,但是只是碰撞的时候发出一次声音,所以引入location if (self.ball.x == 0 || self.ball.x == self.view.width - self.ball.width || self.ball.y == self.view.height - self.ball.height || self.ball.y == 0) { if (self.location.x == self.ball.x ||self.location.y == self.ball.y ){ } else{ //播放声音 [AudioTool playWithFileName:@"1.aif"]; } } //判断 ,如果在边界上了,但是继续横向滚动,碰到了两边的边界,需要发出声音 if (self.ball.y == 0 || self.ball.y == self.view.height - self.ball.height) { if ((self.ball.x == 0) && (self.ball.x != self.location.x)) { [AudioTool playWithFileName:@"1.aif"]; } if ((self.ball.x == self.view.width- self.ball.width) && (self.ball.x != self.location.x)) { [AudioTool playWithFileName:@"1.aif"]; } } if (self.ball.x == 0 || self.ball.x == self.view.width - self.ball.width) { if ((self.ball.y == 0) && (self.ball.y != self.location.y)) { [AudioTool playWithFileName:@"1.aif"]; } if ((self.ball.y == self.view.height- self.ball.height) && (self.ball.y != self.location.y)) { [AudioTool playWithFileName:@"1.aif"]; } } //将此刻的小球位置做记录 _location.x = self.ball.x; _location.y = self.ball.y; }]; } @end
7、摇一摇简单实现
(1)监控摇一摇的方法
方法1:通过分析加速计数据来判断是否进行了摇一摇操作(比较复杂)
方法2:iOS自带的Shake监控API(非常简单,UIKit框架里的)
(2)判断摇一摇的步骤:实现3个摇一摇监听方法
<1> 检测到摇动
- (void)motionBegan:(UIEventSubtype)motion withEvent:(UIEvent *)event
<2> 摇动取消(被中断)
- (void)motionCancelled:(UIEventSubtype)motion withEvent:(UIEvent *)event
<3> 摇动结束
- (void)motionEnded:(UIEventSubtype)motion withEvent:(UIEvent *)event
四、距离传感器
1、位置在听筒上方,常见应用:打电话时,脸部靠近屏幕会自动锁屏,并黑屏。
2、配置示例
#import "ViewController.h" @implementation ViewController - (void)viewDidLoad { [super viewDidLoad]; //打开一个距离传感器的开关 [UIDevice currentDevice].proximityMonitoringEnabled = YES; //添加一个监听通知 [[NSNotificationCenter defaultCenter]addObserver:self selector:@selector(proximityStateDidChange:) name:UIDeviceProximityStateDidChangeNotification object:nil]; } - (void)proximityStateDidChange:(NSNotification *)noti { if ([UIDevice currentDevice].proximityState) { NSLog(@"手机靠近了"); }else{ NSLog(@"手机拿远了!"); } } @end
五、蓝牙(4.0BLE)
1、iOS中蓝牙的实现方案:iOS中提供了4个框架用于实现蓝牙连接
(1)GameKit.framework(用法简单)
只能用于iOS设备之间的连接,多用于游戏(比如五子棋对战),从iOS7开始过期
(2)MultipeerConnectivity.framework
只能用于iOS设备之间的连接,从iOS7开始引入,主要用于文件共享(仅限于沙盒的文件)
(3)ExternalAccessory.framework
可用于第三方蓝牙设备交互,但是蓝牙设备必须经过苹果MFi认证(国内较少)
(4)CoreBluetooth.framework(时下热门)
<1> 可用于第三方蓝牙设备交互,必须要支持蓝牙4.0
<2> 硬件至少是4s,系统至少是iOS6
<3>蓝牙4.0以低功耗著称,一般也叫BLE(Bluetooth Low Energy)
<4>目前应用比较多的案例:运动手坏、嵌入式设备、智能家居
2、Core Bluetooth
(1)核心结构图
(2)蓝牙基本常识
<1> 每个蓝牙4.0设备都是通过服务(Service)和特征(Characteristic)来展示自己的
<2> 一个设备必然包含一个或多个服务,每个服务下面又包含若干个特征
<3> 特征是与外界交互的最小单位:
比如说,一台蓝牙4.0设备,用特征A来描述自己的出厂信息,用特征B来收发数据
<4>服务和特征都是用UUID来唯一标识的,通过UUID就能区别不同的服务和特征
<5>设备里面各个服务(service)和特征(characteristic)的功能,均由蓝牙设备硬件厂商提供,比如哪些是用来交互(读写),哪些可获取模块信息(只读)等
(3)Core Bluetooth的开发步骤
<1>建立中心设备
<2>扫描外设(Discover Peripheral)
<3>连接外设(Connect Peripheral)
<4>扫描外设中的服务和特征(Discover Services And Characteristics)
<5>利用特征与外设做数据交互(Explore And Interact)
<6>断开连接(Disconnect)
3、配置示例(模拟器无法运行)
#import "ViewController.h" #import <CoreBluetooth/CoreBluetooth.h> @interface ViewController ()<CBCentralManagerDelegate,CBPeripheralDelegate> @property (nonatomic, strong) CBCentralManager *centralManager; @property (nonatomic, strong) NSMutableArray *peripherals; @property (nonatomic, strong) CBCharacteristic *characteristic; @end @implementation ViewController - (void)viewDidLoad { [super viewDidLoad]; //创建蓝牙管理类 CBCentralManager *centralManager = [[CBCentralManager alloc]initWithDelegate:self queue:dispatch_get_global_queue(DISPATCH_QUEUE_PRIORITY_DEFAULT, 0)]; self.centralManager = centralManager; } //选择一个设备 开始连接 - (void)btnClickDidselect { CBPeripheral *peripheral = self.peripherals[2]; [self.centralManager connectPeripheral:peripheral options:nil]; } - (void)btnClick { //扫描外设 //uuid 外设的服务或者特征唯一标识,由设备厂商提供,在说明书中 //XXXXXX 服务下面 有可能有很多特征 TTTTTT WWWWWWW CBUUID *uuid = [CBUUID UUIDWithString:@"XXXXXX"]; [self.centralManager scanForPeripheralsWithServices:@[uuid] options:nil]; } - (void)btnClickOnSwitch { //记录之前这个特征所在的外设,然后调用读写方法,进行数据的传输 //外设会根据数据做出相应的反馈 // - (void)writeValue:(NSData *)data forCharacteristic:(CBCharacteristic *)characteristic type:(CBCharacteristicWriteType)type; //self.characteristic } #pragma mark CBCentralManagerDelegate //发现外部设备的时候就会调用 //参数1 中心设备 参数2 外设 - (void)centralManager:(CBCentralManager *)central didDiscoverPeripheral:(CBPeripheral *)peripheral advertisementData:(NSDictionary<NSString *,id> *)advertisementData RSSI:(NSNumber *)RSSI { if (![self.peripherals containsObject:peripheral]) { [self.peripherals addObject:peripheral]; } } //连接外部设备成功就会调用 //参数1 中心设备 参数2 外部设备 - (void)centralManager:(CBCentralManager *)central didConnectPeripheral:(CBPeripheral *)peripheral { //peripheral 扫描服务 或者 特征 [peripheral discoverServices:@[[CBUUID UUIDWithString:@"XXXXXX"]]]; peripheral.delegate = self; } #pragma mark - CBPeripheralDelegate //扫描到外设中的服务的时候就会调用 - (void)peripheral:(CBPeripheral *)peripheral didDiscoverServices:(NSError *)error { for (CBService *service in peripheral.services) { [peripheral discoverCharacteristics:@[[CBUUID UUIDWithString:@"TTTTTT"]] forService:service]; } //例如蓝牙灯泡外设 A 1(开关) 2(亮度) 3(冷暖光) 4(颜色) B 5(能耗) 6(定时) 7 8 } - (void)peripheral:(CBPeripheral *)peripheral didDiscoverCharacteristicsForService:(CBService *)service error:(NSError *)error { for (CBCharacteristic *characteristic in service.characteristics) { if ([characteristic.UUID.UUIDString isEqualToString:@"TTTTTT"]) { self.characteristic = characteristic; } } } @end