• 海康相机打开的方法


        cam = cv2.VideoCapture("rtsp://admin:abcd1234@192.168.31.196/ch1/main/av_stream")

    # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved.
    import argparse
    import cv2, os

    from fcos_core.config import cfg
    from predictor_dc1 import COCODemo

    import time


    def main():
    parser = argparse.ArgumentParser(description="PyTorch Object Detection Webcam Demo")
    parser.add_argument(
    "--config-file",
    default="configs/fcos/fcos_imprv_R_50_FPN_1x.yaml",
    metavar="FILE",
    help="path to config file",
    )
    parser.add_argument(
    "--weights",
    default="FCOS_imprv_R_50_FPN_1x.pth",
    metavar="FILE",
    help="path to the trained model",
    )
    parser.add_argument(
    "--images-dir",
    default="demo/images",
    metavar="DIR",
    help="path to demo images directory",
    )
    parser.add_argument(
    "--min-image-size",
    type=int,
    default=800,
    help="Smallest size of the image to feed to the model. "
    "Model was trained with 800, which gives best results",
    )
    parser.add_argument(
    "opts",
    help="Modify model config options using the command-line",
    default=None,
    nargs=argparse.REMAINDER,
    )

    args = parser.parse_args()

    # load config from file and command-line arguments
    cfg.merge_from_file(args.config_file)
    cfg.merge_from_list(args.opts)
    cfg.MODEL.WEIGHT = args.weights

    cfg.freeze()

    # The following per-class thresholds are computed by maximizing
    # per-class f-measure in their precision-recall curve.
    # Please see compute_thresholds_for_classes() in coco_eval.py for details.
    thresholds_for_classes = [
    0.4923645853996277, 0.4928510785102844, 0.5040897727012634,
    0.4912887513637543, 0.5016880631446838, 0.5278812646865845,
    0.5351834893226624, 0.5003424882888794, 0.4955945909023285,
    0.43564629554748535, 0.6089804172515869, 0.666087806224823,
    0.5932040214538574, 0.48406165838241577, 0.4062422513961792,
    0.5571075081825256, 0.5671307444572449, 0.5268378257751465,
    0.5112953186035156, 0.4647842049598694, 0.5324517488479614,
    0.5795850157737732, 0.5152440071105957, 0.5280804634094238,
    0.4791383445262909, 0.5261335372924805, 0.4906163215637207,
    0.523737907409668, 0.47027698159217834, 0.5103300213813782,
    0.4645252823829651, 0.5384289026260376, 0.47796186804771423,
    0.4403403103351593, 0.5101461410522461, 0.5535093545913696,
    0.48472103476524353, 0.5006796717643738, 0.5485560894012451,
    0.4863888621330261, 0.5061569809913635, 0.5235867500305176,
    0.4745445251464844, 0.4652363359928131, 0.4162440598011017,
    0.5252017974853516, 0.42710989713668823, 0.4550687372684479,
    0.4943239390850067, 0.4810051918029785, 0.47629663348197937,
    0.46629616618156433, 0.4662836790084839, 0.4854755401611328,
    0.4156557023525238, 0.4763634502887726, 0.4724511504173279,
    0.4915047585964203, 0.5006274580955505, 0.5124194622039795,
    0.47004589438438416, 0.5374764204025269, 0.5876904129981995,
    0.49395060539245605, 0.5102297067642212, 0.46571290493011475,
    0.5164387822151184, 0.540651798248291, 0.5323763489723206,
    0.5048757195472717, 0.5302401781082153, 0.48333442211151123,
    0.5109739303588867, 0.4077408015727997, 0.5764586925506592,
    0.5109297037124634, 0.4685552418231964, 0.5148998498916626,
    0.4224434792995453, 0.4998510777950287
    ]

    demo_im_names = os.listdir(args.images_dir)

    # prepare object that handles inference plus adds predictions on top of image
    coco_demo = COCODemo(
    cfg,
    confidence_thresholds_for_classes=thresholds_for_classes,
    min_image_size=args.min_image_size
    )


    video_path="/data1/duchao/FCOS/demo/vidios/20201028_20201028140638_20201028141631_1.mp4"
    save_dir="/data1/duchao/FCOS/demo/vidios/result"
    write_video_name = os.path.basename(video_path).split(".")[0]
    #cam = cv2.VideoCapture(video_path)
    cam = cv2.VideoCapture("rtsp://admin:abcd1234@192.168.31.196/ch1/main/av_stream")

    count = 0
    fps = cam.get(cv2.CAP_PROP_FPS)
    #fps=30
    size = (int(cam.get(cv2.CAP_PROP_FRAME_WIDTH)), int(cam.get(cv2.CAP_PROP_FRAME_HEIGHT)))
    fourcc = cv2.VideoWriter_fourcc('M', 'J', 'P', 'G')
    print("fps: {}".format(fps)) #可能是错的,仅做参考
    print("size: {}".format(size))
    video_writer = cv2.VideoWriter(save_dir+"/"+write_video_name+".avi", fourcc, fps, size)
    print("video is ",video_path,save_dir+"/"+write_video_name+".avi")

    while True:
    start_time = time.time()
    ret_val, img = cam.read()
    print(ret_val)
    if img is None:
    continue
    start_time = time.time()
    composite = coco_demo.run_on_opencv_image(img)
    print("{} inference time: {:.2f}s".format("im_name", time.time() - start_time))
    print("write")
    video_writer.write(composite)
    #cv2.imwrite("/data1/duchao/FCOS/demo/images/RESULT/"+im_name, composite)
    print("Press any keys to exit ...")
    #cv2.waitKey()
    #cv2.destroyAllWindows()

    if __name__ == "__main__":
    main()

  • 相关阅读:
    软工实践第三次作业(结对第一次作业)
    软工实践第四次作业(团队展示)
    软工实践第二次作业(词频统计)
    软工实践第七次作业(软件产品案例分析 )
    安装cordova和ionic遇到cordova v和ionic v出错的问题,出现SyntaxError
    软工实践第八次作业(软件工程实践总结)
    软工实践第六次作业(团队项目作业汇总)
    软工实践第五次作业(结对第二次作业)
    chain of responsibilit职责链模式
    术语
  • 原文地址:https://www.cnblogs.com/shuimuqingyang/p/14462886.html
Copyright © 2020-2023  润新知