Browse Source

字幕服务分开初始化

jsonwang 3 năm trước cách đây
mục cha
commit
6049b94c0e

+ 115 - 106
BFRecordScreenKit/Classes/RecordScreen/Controller/BFRecordScreenController.swift

@@ -127,6 +127,10 @@ public class BFRecordScreenController: BFBaseViewController {
 
     var indirectionView: BFIndirectionProgressView?
 
+    //字幕识别 SDK 初始化参数
+    var NeoNuiToken:String?
+    var NeoNuiAppid:String?
+     
     lazy var progreddL: UILabel = {
         let l = UILabel()
         l.textAlignment = .center
@@ -399,135 +403,138 @@ public class BFRecordScreenController: BFBaseViewController {
         // add by ak 取 nsl token
         BFRecordScreenViewModel.getNlsAccessToken { [weak self] token, appkey in
             BFLog(message: "nls appkey is \(appkey), token is \(token)")
-
-            self?.recorderManager = BFVoiceRecordManager(token: token, appid: appkey)
-
-            // 录音进度
-            self?.recorderManager?.recorderProgrossHandle = { [weak self] progress in
-                BFLog(1, message: "curr:录音进度--\(progress) \(self?.recordStartTime ?? 0) \(self?.isRecording ?? false)")
-                self?.drawProgressIndication(progress: progress ?? 0)
+            self?.NeoNuiAppid = appkey
+            self?.NeoNuiToken = token
+        
+        }
+        
+        //录音字幕识别初始化相关
+        recorderManager = BFVoiceRecordManager()
+        // 录音进度
+        recorderManager?.recorderProgrossHandle = { [weak self] progress in
+            BFLog(1, message: "curr:录音进度--\(progress) \(self?.recordStartTime ?? 0) \(self?.isRecording ?? false)")
+            self?.drawProgressIndication(progress: progress ?? 0)
+        }
+
+        // 录音字幕回调
+        recorderManager?.subtitleRecordHandle = { [weak self] asrResult, audioFilePath in
+            if asrResult == nil {
+                BFLog(message: "识别结果为空????不能生成字幕数据")
+                return
             }
+            let dicResult: [String: Any]? = jsonStringToDictionary(asrResult!)
 
-            // 录音字幕回调
-            self?.recorderManager?.subtitleRecordHandle = { [weak self] asrResult, audioFilePath in
-                if asrResult == nil {
-                    BFLog(message: "识别结果为空????不能生成字幕数据")
-                    return
-                }
-                let dicResult: [String: Any]? = jsonStringToDictionary(asrResult!)
-
-                let header = dicResult?["header"] as? [String: Any]
-                let payload = dicResult?["payload"] as? [String: Any]
+            let header = dicResult?["header"] as? [String: Any]
+            let payload = dicResult?["payload"] as? [String: Any]
 
-                BFLog(message: "识别结果:) \((payload?["result"])!) startTime:\(self?.recorderManager?.voiceModel?.startTime ?? 0.0)")
-                DispatchQueue.main.async {
-                    // 1,保存字幕数据 begin_time是开始出现文字的时间,time 是结束文字出现的时间 单位都为毫秒,都是相对于录制音频数据整段时间。self.recorderManager.voiceModel?.startTime 为开始的录制的时间,开始和结束都要加上这个时差
+            BFLog(message: "识别结果:) \((payload?["result"])!) startTime:\(self?.recorderManager?.voiceModel?.startTime ?? 0.0)")
+            DispatchQueue.main.async {
+                // 1,保存字幕数据 begin_time是开始出现文字的时间,time 是结束文字出现的时间 单位都为毫秒,都是相对于录制音频数据整段时间。self.recorderManager.voiceModel?.startTime 为开始的录制的时间,开始和结束都要加上这个时差
 
-                    let newSubtitle = PQEditSubTitleModel()
-                    // 任务全局唯一ID,请记录该值,便于排查问题。 每次 startRecorder 和 stopRecoder 之间  task_Id都不会变化
-                    newSubtitle.taskID = (header?["task_id"] as? String) ?? ""
+                let newSubtitle = PQEditSubTitleModel()
+                // 任务全局唯一ID,请记录该值,便于排查问题。 每次 startRecorder 和 stopRecoder 之间  task_Id都不会变化
+                newSubtitle.taskID = (header?["task_id"] as? String) ?? ""
 
-                    // 这里加300ms 是因为返回结果为了切到字,时长提前一些时间,具体时间官方没说和原音频有关系。这里我们先延后300ms 单位:毫秒。
+                // 这里加300ms 是因为返回结果为了切到字,时长提前一些时间,具体时间官方没说和原音频有关系。这里我们先延后300ms 单位:毫秒。
 
-                    newSubtitle.timelineIn = (self?.recorderManager?.voiceModel?.startTime ?? 0.0) + Float64((((payload?["begin_time"]) as? Int) ?? 0) + 300) / 1000.0
+                newSubtitle.timelineIn = (self?.recorderManager?.voiceModel?.startTime ?? 0.0) + Float64((((payload?["begin_time"]) as? Int) ?? 0) + 300) / 1000.0
 
-                    newSubtitle.timelineOut = (self?.recorderManager?.voiceModel?.startTime ?? 0.0) + Float64(((payload?["time"]) as? Int) ?? 0) / 1000.0
-                    var showText = ((payload?["result"]) as? String) ?? ""
-                    if showText.count > subtitleMaxlength {
-                        showText = showText.substring(to: subtitleMaxlength)
-                        showText += "..."
-                    }
-                    newSubtitle.text = showText
-                    newSubtitle.audioFilePath = audioFilePath ?? ""
+                newSubtitle.timelineOut = (self?.recorderManager?.voiceModel?.startTime ?? 0.0) + Float64(((payload?["time"]) as? Int) ?? 0) / 1000.0
+                var showText = ((payload?["result"]) as? String) ?? ""
+                if showText.count > subtitleMaxlength {
+                    showText = showText.substring(to: subtitleMaxlength)
+                    showText += "..."
+                }
+                newSubtitle.text = showText
+                newSubtitle.audioFilePath = audioFilePath ?? ""
 
-                    BFLog(message: "添加字幕数据 timelineIn \(newSubtitle.timelineIn) timelineOut \(newSubtitle.timelineOut) text: \(newSubtitle.text) 音频路径为:\(audioFilePath ?? "")")
-                    newSubtitle.setting = self?.subtitleSettingView.subtitle.setting ?? BFSubTitileSettingModel()
+                BFLog(message: "添加字幕数据 timelineIn \(newSubtitle.timelineIn) timelineOut \(newSubtitle.timelineOut) text: \(newSubtitle.text) 音频路径为:\(audioFilePath ?? "")")
+                newSubtitle.setting = self?.subtitleSettingView.subtitle.setting ?? BFSubTitileSettingModel()
 
-                    self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.append(newSubtitle)
-                }
+                self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.append(newSubtitle)
             }
-            // 录音结束
-            self?.recorderManager?.endRecordHandle = { [weak self] voideModel, _ in
-                if let sself = self, let model = voideModel, FileManager.default.fileExists(atPath: model.wavFilePath ?? "") {
-                    // 加入到语音数组里
-                    model.endTime = sself.currentAssetProgress.seconds
+        }
+        // 录音结束
+        recorderManager?.endRecordHandle = { [weak self] voideModel, _ in
+            if let sself = self, let model = voideModel, FileManager.default.fileExists(atPath: model.wavFilePath ?? "") {
+                // 加入到语音数组里
+                model.endTime = sself.currentAssetProgress.seconds
 
-                    let newRange = CMTimeRange(start: CMTime(seconds: model.startTime, preferredTimescale: 1000), end: CMTime(seconds: model.endTime, preferredTimescale: 1000))
+                let newRange = CMTimeRange(start: CMTime(seconds: model.startTime, preferredTimescale: 1000), end: CMTime(seconds: model.endTime, preferredTimescale: 1000))
 
-                    var deletedVoices = [(PQVoiceModel, Int)]()
+                var deletedVoices = [(PQVoiceModel, Int)]()
 
-                    for (i, m) in sself.itemModels[sself.currItemModelIndex].voiceStickers.enumerated() {
-                        let originRange = CMTimeRange(start: CMTime(seconds: m.startTime, preferredTimescale: 1000), end: CMTime(seconds: m.endTime, preferredTimescale: 1000))
+                for (i, m) in sself.itemModels[sself.currItemModelIndex].voiceStickers.enumerated() {
+                    let originRange = CMTimeRange(start: CMTime(seconds: m.startTime, preferredTimescale: 1000), end: CMTime(seconds: m.endTime, preferredTimescale: 1000))
 
-                        if CMTimeRangeGetIntersection(originRange, otherRange: newRange).duration.seconds > 0 {
-                            deletedVoices.append((m, i))
-                            continue
-                        }
-                    }
-                    sself.itemModels[sself.currItemModelIndex].voiceStickers.removeAll { m in
-                        let originRange = CMTimeRange(start: CMTime(seconds: m.startTime, preferredTimescale: 1000), end: CMTime(seconds: m.endTime, preferredTimescale: 1000))
-                        return CMTimeRangeGetIntersection(originRange, otherRange: newRange).duration.seconds > 0
+                    if CMTimeRangeGetIntersection(originRange, otherRange: newRange).duration.seconds > 0 {
+                        deletedVoices.append((m, i))
+                        continue
                     }
-                    BFLog(1, message: "添加录音文件:\(model.startTime) -- \(model.endTime)")
+                }
+                sself.itemModels[sself.currItemModelIndex].voiceStickers.removeAll { m in
+                    let originRange = CMTimeRange(start: CMTime(seconds: m.startTime, preferredTimescale: 1000), end: CMTime(seconds: m.endTime, preferredTimescale: 1000))
+                    return CMTimeRangeGetIntersection(originRange, otherRange: newRange).duration.seconds > 0
+                }
+                BFLog(1, message: "添加录音文件:\(model.startTime) -- \(model.endTime)")
 
-                    var event = sself.events.last
-                    if event != nil {
-                        event!.deletedVoices = deletedVoices
-                        sself.events.removeLast()
-                        sself.events.append(event!)
-                    }
-                    sself.itemModels[sself.currItemModelIndex].voiceStickers.append(model)
-                    if sself.itemModels[sself.currItemModelIndex].mediaType == .IMAGE {
-                        var duration: Double = 0
-                        sself.itemModels[sself.currItemModelIndex].voiceStickers.forEach { temp in
-                            temp.duration = "\(temp.endTime - temp.startTime)"
-                            duration = duration + (temp.endTime - temp.startTime)
-                        }
-                        sself.itemModels[sself.currItemModelIndex].materialDuraion = Double(String(format: "%.3f", duration)) ?? 0
-                        self?.isEndPlay = true
-                        // 录制结束显示播放按钮
-                        (sself.collectionView.cellForItem(at: IndexPath(item: sself.currItemModelIndex, section: 0)) as? BFImageCoverViewCell)?.playBtn.isSelected = sself.itemModels[sself.currItemModelIndex].voiceStickers.count <= 0
-                    }
-                    DispatchQueue.main.async { [weak self] in
-                        // 录音完,重绘撤销按钮,更新录音按钮,
-                        self?.changeWithDrawBtnLayout(true)
-                        // 注:在录制结束时矫正当前位置,避免跟指针无法对其
-                        self?.indirectionView?.resetCurrentItem(start: model.startTime, end: model.endTime)
-                        // 矫正进度
-                        self?.resetCurrentProgress()
-                        self?.deleteRecordBtn.isHidden = true
-                        self?.recordBtn.isHidden = false
+                var event = sself.events.last
+                if event != nil {
+                    event!.deletedVoices = deletedVoices
+                    sself.events.removeLast()
+                    sself.events.append(event!)
+                }
+                sself.itemModels[sself.currItemModelIndex].voiceStickers.append(model)
+                if sself.itemModels[sself.currItemModelIndex].mediaType == .IMAGE {
+                    var duration: Double = 0
+                    sself.itemModels[sself.currItemModelIndex].voiceStickers.forEach { temp in
+                        temp.duration = "\(temp.endTime - temp.startTime)"
+                        duration = duration + (temp.endTime - temp.startTime)
                     }
-                    sself.currentPlayRecordIndex = -3 // 刚录音完,不需要播放录音
-                    // 重置录制开始时间
-                    sself.recordStartTime = 0
+                    sself.itemModels[sself.currItemModelIndex].materialDuraion = Double(String(format: "%.3f", duration)) ?? 0
+                    self?.isEndPlay = true
+                    // 录制结束显示播放按钮
+                    (sself.collectionView.cellForItem(at: IndexPath(item: sself.currItemModelIndex, section: 0)) as? BFImageCoverViewCell)?.playBtn.isSelected = sself.itemModels[sself.currItemModelIndex].voiceStickers.count <= 0
+                }
+                DispatchQueue.main.async { [weak self] in
+                    // 录音完,重绘撤销按钮,更新录音按钮,
+                    self?.changeWithDrawBtnLayout(true)
+                    // 注:在录制结束时矫正当前位置,避免跟指针无法对其
+                    self?.indirectionView?.resetCurrentItem(start: model.startTime, end: model.endTime)
+                    // 矫正进度
+                    self?.resetCurrentProgress()
+                    self?.deleteRecordBtn.isHidden = true
+                    self?.recordBtn.isHidden = false
                 }
+                sself.currentPlayRecordIndex = -3 // 刚录音完,不需要播放录音
+                // 重置录制开始时间
+                sself.recordStartTime = 0
             }
-            self?.recorderManager?.cancelRecordHandle = { [weak self] voiceModel in
-                // add by ak 取消录制后删除对应字幕数据,这里可恢复操作吗?
-                var subtitleCount = self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.count ?? 0
-                BFLog(2, message: "删除\(voiceModel?.wavFilePath ?? "")对应的字幕 前 count\(subtitleCount)")
-                if subtitleCount > 0 {
-                    for title in self!.itemModels[self?.currItemModelIndex ?? 0].titleStickers {
-                        if title.audioFilePath == voiceModel?.wavFilePath ?? "" {
-                            if let index = self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.firstIndex(of: title) {
-                                self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.remove(at: index)
-                            }
+        }
+        recorderManager?.cancelRecordHandle = { [weak self] voiceModel in
+            // add by ak 取消录制后删除对应字幕数据,这里可恢复操作吗?
+            var subtitleCount = self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.count ?? 0
+            BFLog(2, message: "删除\(voiceModel?.wavFilePath ?? "")对应的字幕 前 count\(subtitleCount)")
+            if subtitleCount > 0 {
+                for title in self!.itemModels[self?.currItemModelIndex ?? 0].titleStickers {
+                    if title.audioFilePath == voiceModel?.wavFilePath ?? "" {
+                        if let index = self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.firstIndex(of: title) {
+                            self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.remove(at: index)
                         }
                     }
                 }
-                subtitleCount = self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.count ?? 0
-                BFLog(2, message: "删除\(voiceModel?.wavFilePath ?? "")对应的字幕  后 count\(subtitleCount)")
-
-                // 取消录制以后重置进度
-                self?.withdrawAction()
-                // 重置录制开始时间
-                self?.recordStartTime = 0
-                /// 重置进度
-                self?.resetCurrentProgress()
-                // 移除
-                self?.indirectionView?.deleteItem(isCurrent: true)
             }
+            subtitleCount = self?.itemModels[self?.currItemModelIndex ?? 0].titleStickers.count ?? 0
+            BFLog(2, message: "删除\(voiceModel?.wavFilePath ?? "")对应的字幕  后 count\(subtitleCount)")
+
+            // 取消录制以后重置进度
+            self?.withdrawAction()
+            // 重置录制开始时间
+            self?.recordStartTime = 0
+            /// 重置进度
+            self?.resetCurrentProgress()
+            // 移除
+            self?.indirectionView?.deleteItem(isCurrent: true)
         }
 
         view.backgroundColor = .black
@@ -883,6 +890,8 @@ public class BFRecordScreenController: BFBaseViewController {
         model.volume = 100
         recorderManager?.voiceModel = model
         recorderManager?.startRecord()
+        //开始字幕识别
+        recorderManager?.audioRecorder?.startNeoNui((NeoNuiToken ?? ""), appid: (NeoNuiAppid ?? ""))
         if recordStartTime <= 0 {
             recordStartTime = currentAssetProgress.seconds
         }