Android实现屏幕录制功能

这篇文章主要为大家详细介绍了Android实现屏幕录制功能,文中示例代码介绍的非常详细,具有一定的参考价值,感兴趣的小伙伴们可以参考一下

本文实例为大家分享了Android实现屏幕录制功能的具体代码,供大家参考,具体内容如下

1.效果图:

2.添加依赖 


dependencies {
 implementation fileTree(dir: 'libs', include: ['*.jar'])
 implementation"org.jetbrains.kotlin:kotlin-stdlib-jdk7:$kotlin_version"
 implementation 'androidx.appcompat:appcompat:1.1.0'
 implementation 'androidx.core:core-ktx:1.0.2'
 implementation 'androidx.constraintlayout:constraintlayout:1.1.3'
 testImplementation 'junit:junit:4.12'
 androidTestImplementation 'androidx.test.ext:junit:1.1.1'
 androidTestImplementation 'androidx.test.espresso:espresso-core:3.2.0'
 api 'com.blankj:utilcode:1.24.4'
}
repositories {
 mavenCentral()
}

3.注册权限:


<uses-permission android:name="android.permission.RECORD_AUDIO" />
<uses-permission android:name="android.permission.WRITE_EXTERNAL_STORAGE" />
<uses-permission android:name="android.permission.READ_EXTERNAL_STORAGE" />

4.主界面,

test.aac是录屏的时候配的音乐,可以随便找另外一个放到assets文件夹里面进行替换


package com.ufi.pdioms.ztkotlin
 
 
import android.content.Intent
import android.content.res.AssetFileDescriptor
import android.media.MediaPlayer
import android.os.Build
import androidx.appcompat.app.AppCompatActivity
import android.os.Bundle
import android.util.Log
import android.widget.Toast
import com.blankj.utilcode.util.PathUtils
import kotlinx.android.synthetic.main.activity_main.*
 
class MainActivity : AppCompatActivity() {
 // https://github.com/fanqilongmoli/AndroidScreenRecord
 private var screenRecordHelper: ScreenRecordHelper? = null
 private val afdd:AssetFileDescriptor by lazy { assets.openFd("test.aac") }
 
 override fun onCreate(savedInstanceState: Bundle?) {
  super.onCreate(savedInstanceState)
  setContentView(R.layout.activity_main)
 
  btnStart.setOnClickListener {
   if (Build.VERSION.SDK_INT >= Build.VERSION_CODES.LOLLIPOP) {
    if (screenRecordHelper == null) {
     screenRecordHelper = ScreenRecordHelper(this, object : ScreenRecordHelper.OnVideoRecordListener {
      override fun onBeforeRecord() {
      }
 
      override fun onStartRecord() {
       play()
      }
 
      override fun onCancelRecord() {
       releasePlayer()
      }
 
      override fun onEndRecord() {
       releasePlayer()
      }
 
     }, PathUtils.getExternalStoragePath() + "/fanqilong")
    }
    screenRecordHelper?.apply {
     if (!isRecording) {
      // 如果你想录制音频(一定会有环境音量),你可以打开下面这个限制,并且使用不带参数的 stopRecord()
//      recordAudio = true
      startRecord()
     }
    }
   } else {
    Toast.makeText(this@MainActivity.applicationContext, "sorry,your phone does not support recording screen", Toast.LENGTH_LONG).show()
   }
  }
 
  btnStop.setOnClickListener {
   if (Build.VERSION.SDK_INT >= Build.VERSION_CODES.LOLLIPOP) {
    screenRecordHelper?.apply {
     if (isRecording) {
      if (mediaPlayer != null) {
       // 如果选择带参数的 stop 方法,则录制音频无效
       stopRecord(mediaPlayer!!.duration.toLong(), 15 * 1000, afdd)
      } else {
       stopRecord()
      }
     }
    }
   }
  }
 }
 
 private fun play() {
  mediaPlayer = MediaPlayer()
  try {
   mediaPlayer?.apply {
    this.reset()
    this.setDataSource(afdd.fileDescriptor, afdd.startOffset, afdd.length)
    this.isLooping = true
    this.prepare()
    this.start()
   }
  } catch (e: Exception) {
   Log.d("fanqilong", "播放音乐失败")
  } finally {
 
  }
 }
 
 // 音频播放
 private var mediaPlayer: MediaPlayer? = null
 
 private fun releasePlayer() {
  mediaPlayer?.apply {
   stop()
   release()
  }
  mediaPlayer = null
 }
 
 override fun onActivityResult(requestCode: Int, resultCode: Int, data: Intent?) {
  super.onActivityResult(requestCode, resultCode, data)
  if (Build.VERSION.SDK_INT >= Build.VERSION_CODES.LOLLIPOP && data != null) {
   screenRecordHelper?.onActivityResult(requestCode, resultCode, data)
  }
 }
 
 override fun onDestroy() {
  if (Build.VERSION.SDK_INT >= Build.VERSION_CODES.LOLLIPOP) {
   screenRecordHelper?.clearAll()
  }
  afdd.close()
  super.onDestroy()
 }
}

5.录屏代码


package com.ufi.pdioms.ztkotlin
 
import android.app.Activity
import android.content.Context
import android.content.Intent
import android.content.pm.PackageManager
import android.content.res.AssetFileDescriptor
import android.hardware.display.DisplayManager
import android.hardware.display.VirtualDisplay
import android.media.*
import android.media.projection.MediaProjection
import android.media.projection.MediaProjectionManager
import android.net.Uri
import android.os.Build
import android.os.Environment
import android.os.Handler
import android.util.DisplayMetrics
import android.util.Log
import android.widget.Toast
import androidx.annotation.RequiresApi
import com.blankj.utilcode.constant.PermissionConstants
import com.blankj.utilcode.util.PermissionUtils
import java.io.File
import java.lang.Exception
import java.nio.ByteBuffer
 
@RequiresApi(Build.VERSION_CODES.LOLLIPOP)
class ScreenRecordHelper @JvmOverloads constructor(
 private var activity: Activity,
 private val listener: OnVideoRecordListener?,
 private var savePath: String = Environment.getExternalStorageDirectory().absolutePath + File.separator
   + "DCIM" + File.separator + "Camera",
 private val saveName: String = "record_${System.currentTimeMillis()}"
) {
 
 private val mediaProjectionManager by lazy { activity.getSystemService(Context.MEDIA_PROJECTION_SERVICE) as? MediaProjectionManager }
 private var mediaRecorder: MediaRecorder? = null
 private var mediaProjection: MediaProjection? = null
 private var virtualDisplay: VirtualDisplay? = null
 private val displayMetrics by lazy { DisplayMetrics() }
 private var saveFile: File? = null
 var isRecording = false
 var recordAudio = false
 
 init {
  activity.windowManager.defaultDisplay.getMetrics(displayMetrics)
 }
 
 companion object {
  private const val VIDEO_FRAME_RATE = 30
  private const val REQUEST_CODE = 1024
  private const val TAG = "ScreenRecordHelper"
 }
 
 fun startRecord() {
  if (mediaProjectionManager == null) {
   Log.d(TAG, "mediaProjectionManager == null,当前手机暂不支持录屏")
   showToast(R.string.phone_not_support_screen_record)
   return
  }
 
  PermissionUtils.permission(PermissionConstants.STORAGE, PermissionConstants.MICROPHONE)
   .callback(object : PermissionUtils.SimpleCallback {
    override fun onGranted() {
     mediaProjectionManager?.apply {
      listener?.onBeforeRecord()
      val intent = this.createScreenCaptureIntent()
      if (activity.packageManager.resolveActivity(
        intent,
        PackageManager.MATCH_DEFAULT_ONLY
       ) != null
      ) {
       activity.startActivityForResult(intent, REQUEST_CODE)
      } else {
       showToast(R.string.phone_not_support_screen_record)
      }
     }
    }
 
    override fun onDenied() {
     showToast(R.string.permission_denied)
    }
 
   }).request()
 }
 
 @RequiresApi(Build.VERSION_CODES.N)
 fun resume() {
  mediaRecorder?.resume()
 }
 
 @RequiresApi(Build.VERSION_CODES.N)
 fun pause() {
  mediaRecorder?.pause()
 }
 
 fun onActivityResult(requestCode: Int, resultCode: Int, data: Intent) {
  if (requestCode == REQUEST_CODE) {
   if (resultCode == Activity.RESULT_OK) {
    mediaProjection = mediaProjectionManager!!.getMediaProjection(resultCode, data)
 
    // 部分手机录制视频的时候 会出现弹框
    Handler().postDelayed({
     if (initRecorder()) {
      isRecording = true
      mediaRecorder?.start()
      listener?.onStartRecord()
     } else {
      showToast(R.string.phone_not_support_screen_record)
     }
    }, 150)
   } else {
    showToast(R.string.phone_not_support_screen_record)
   }
  }
 }
 
 fun cancelRecord(){
  stopRecord()
  saveFile?.delete()
  saveFile = null
  listener?.onCancelRecord()
 }
 
 
 fun stopRecord(videoDuration: Long = 0, audioDuration: Long = 0, afdd: AssetFileDescriptor? = null){
  stop()
  if (audioDuration != 0L && afdd != null) {
   syntheticAudio(videoDuration, audioDuration, afdd)
  } else {
   // saveFile
   if (saveFile != null) {
    val newFile = File(savePath, "$saveName.mp4")
    // 录制结束后修改后缀为 mp4
    saveFile!!.renameTo(newFile)
    refreshVideo(newFile)
   }
   saveFile = null
  }
 }
 
 
 private fun refreshVideo(newFile: File) {
  Log.d(TAG, "screen record end,file length:${newFile.length()}.")
  if (newFile.length() > 5000) {
   val intent = Intent(Intent.ACTION_MEDIA_SCANNER_SCAN_FILE)
   intent.data = Uri.fromFile(newFile)
   activity.sendBroadcast(intent)
   Log.e("TAG","refreshVideo: "+savePath)
   showToast(R.string.save_to_album_success)
  } else {
   newFile.delete()
   showToast(R.string.phone_not_support_screen_record)
   Log.d(TAG, activity.getString(R.string.record_faild))
  }
 }
 
 private fun stop() {
  if (isRecording) {
   isRecording = false
   try {
    mediaRecorder?.apply {
     setOnErrorListener(null)
     setOnInfoListener(null)
     setPreviewDisplay(null)
     stop()
     Log.d(TAG, "stop success")
    }
   } catch (e: Exception) {
    Log.e(TAG, "stopRecorder() error!${e.message}")
   } finally {
    mediaRecorder?.reset()
    virtualDisplay?.release()
    mediaProjection?.stop()
    listener?.onEndRecord()
   }
 
 
  }
 }
 
 private fun initRecorder(): Boolean {
  var result = true
  val f = File(savePath)
  if (!f.exists()) {
   f.mkdir()
  }
  saveFile = File(savePath, "$saveName.tmp")
  saveFile?.apply {
   if (exists()) {
    delete()
   }
  }
  mediaRecorder = MediaRecorder()
  val width = Math.min(displayMetrics.widthPixels, 1080)
  val height = Math.min(displayMetrics.heightPixels, 1920)
  mediaRecorder?.apply {
   if (recordAudio) {
    setAudioSource(MediaRecorder.AudioSource.MIC)
   }
   setVideoSource(MediaRecorder.VideoSource.SURFACE)
   setOutputFormat(MediaRecorder.OutputFormat.MPEG_4)
   setVideoEncoder(MediaRecorder.VideoEncoder.H264)
   if (recordAudio) {
    setAudioEncoder(MediaRecorder.AudioEncoder.AMR_NB)
   }
   setOutputFile(saveFile!!.absolutePath)
   setVideoSize(width, height)
   setVideoEncodingBitRate(8388608)
   setVideoFrameRate(VIDEO_FRAME_RATE)
 
   try {
 
    prepare()
    virtualDisplay = mediaProjection?.createVirtualDisplay(
     "MainScreen", width, height, displayMetrics.densityDpi,
     DisplayManager.VIRTUAL_DISPLAY_FLAG_AUTO_MIRROR, surface, null, null
    )
    Log.d(TAG, "initRecorder 成功")
   } catch (e: Exception) {
    Log.e(TAG, "IllegalStateException preparing MediaRecorder: ${e.message}")
    e.printStackTrace()
    result = false
   }
  }
 
  return result
 }
 
 
 private fun showToast(resId: Int) {
  Toast.makeText(activity.applicationContext, activity.applicationContext.getString(resId), Toast.LENGTH_SHORT)
   .show()
 }
 
 fun clearAll() {
  mediaRecorder?.release()
  mediaRecorder = null
  virtualDisplay?.release()
  virtualDisplay = null
  mediaProjection?.stop()
  mediaProjection = null
 }
 
 /**
  * https://stackoverflow.com/questions/31572067/android-how-to-mux-audio-file-and-video-file
  */
 private fun syntheticAudio(audioDuration: Long, videoDuration: Long, afdd: AssetFileDescriptor) {
  Log.d(TAG, "start syntheticAudio")
  val newFile = File(savePath, "$saveName.mp4")
  if (newFile.exists()) {
   newFile.delete()
  }
  try {
   newFile.createNewFile()
   val videoExtractor = MediaExtractor()
   videoExtractor.setDataSource(saveFile!!.absolutePath)
   val audioExtractor = MediaExtractor()
   afdd.apply {
    audioExtractor.setDataSource(fileDescriptor, startOffset, length * videoDuration / audioDuration)
   }
   val muxer = MediaMuxer(newFile.absolutePath, MediaMuxer.OutputFormat.MUXER_OUTPUT_MPEG_4)
   videoExtractor.selectTrack(0)
   val videoFormat = videoExtractor.getTrackFormat(0)
   val videoTrack = muxer.addTrack(videoFormat)
 
   audioExtractor.selectTrack(0)
   val audioFormat = audioExtractor.getTrackFormat(0)
   val audioTrack = muxer.addTrack(audioFormat)
 
   var sawEOS = false
   var frameCount = 0
   val offset = 100
   val sampleSize = 1000 * 1024
   val videoBuf = ByteBuffer.allocate(sampleSize)
   val audioBuf = ByteBuffer.allocate(sampleSize)
   val videoBufferInfo = MediaCodec.BufferInfo()
   val audioBufferInfo = MediaCodec.BufferInfo()
 
   videoExtractor.seekTo(0, MediaExtractor.SEEK_TO_CLOSEST_SYNC)
   audioExtractor.seekTo(0, MediaExtractor.SEEK_TO_CLOSEST_SYNC)
 
   muxer.start()
 
   // 每秒多少帧
   // 实测 OPPO R9em 垃圾手机,拿出来的没有 MediaFormat.KEY_FRAME_RATE
   val frameRate = if (videoFormat.containsKey(MediaFormat.KEY_FRAME_RATE)) {
    videoFormat.getInteger(MediaFormat.KEY_FRAME_RATE)
   } else {
    31
   }
   // 得出平均每一帧间隔多少微妙
   val videoSampleTime = 1000 * 1000 / frameRate
   while (!sawEOS) {
    videoBufferInfo.offset = offset
    videoBufferInfo.size = videoExtractor.readSampleData(videoBuf, offset)
    if (videoBufferInfo.size < 0) {
     sawEOS = true
     videoBufferInfo.size = 0
    } else {
     videoBufferInfo.presentationTimeUs += videoSampleTime
     videoBufferInfo.flags = videoExtractor.sampleFlags
     muxer.writeSampleData(videoTrack, videoBuf, videoBufferInfo)
     videoExtractor.advance()
     frameCount++
    }
   }
   var sawEOS2 = false
   var frameCount2 = 0
   while (!sawEOS2) {
    frameCount2++
    audioBufferInfo.offset = offset
    audioBufferInfo.size = audioExtractor.readSampleData(audioBuf, offset)
 
    if (audioBufferInfo.size < 0) {
     sawEOS2 = true
     audioBufferInfo.size = 0
    } else {
     audioBufferInfo.presentationTimeUs = audioExtractor.sampleTime
     audioBufferInfo.flags = audioExtractor.sampleFlags
     muxer.writeSampleData(audioTrack, audioBuf, audioBufferInfo)
     audioExtractor.advance()
    }
   }
   muxer.stop()
   muxer.release()
   videoExtractor.release()
   audioExtractor.release()
 
   // 删除无声视频文件
   saveFile?.delete()
  } catch (e: Exception) {
   Log.e(TAG, "Mixer Error:${e.message}")
   // 视频添加音频合成失败,直接保存视频
   saveFile?.renameTo(newFile)
 
  } finally {
   afdd.close()
   Handler().post {
    refreshVideo(newFile)
    saveFile = null
   }
  }
 }
 
 
 interface OnVideoRecordListener {
 
  /**
   * 录制开始时隐藏不必要的UI
   */
  fun onBeforeRecord()
 
  /**
   * 开始录制
   */
  fun onStartRecord()
 
  /**
   * 取消录制
   */
  fun onCancelRecord()
 
  /**
   * 结束录制
   */
  fun onEndRecord()
 }
}

6.布局


<?xml version="1.0" encoding="utf-8"?>
<LinearLayout xmlns:android="http://schemas.android.com/apk/res/android"
 xmlns:app="http://schemas.android.com/apk/res-auto"
 xmlns:tools="http://schemas.android.com/tools"
 android:layout_width="match_parent"
 android:layout_height="match_parent"
 android:orientation="vertical"
 tools:context=".MainActivity">
 
 <Button android:id="@+id/btnStart"
  android:layout_width="match_parent"
  android:layout_height="wrap_content"
  android:textAllCaps="false"
  android:text="start"/>
 
 <Button android:id="@+id/btnStop"
  android:layout_width="match_parent"
  android:layout_height="wrap_content"
  android:textAllCaps="false"
  android:text="stop"/>
 
 
</LinearLayout>

以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持得得之家。

本站部分内容来源互联网,如果有图片或者内容侵犯您的权益请联系我们删除!

相关文档推荐

UIScrollView是一个非常重要的控件,其可以展示比设备屏幕更大区域的内容,我们可以通过手指滑动来查看内容视图的每一部分内容,也可以通过手指捏合来对内容视图进行缩放操作,我们每天开发中都不断显式或隐式地与UIScrollView打交道,下面给大家
iOS 和 Android 哪个更利于赚钱?这篇文章为大家揭晓答案,感兴趣的小伙伴们可以参考一下
这篇文章主要介绍了Android Studio 3.5格式化布局代码时错位、错乱bug的解决,具有很好的参考价值,希望对大家有所帮助。一起跟随小编过来看看吧
这篇文章主要介绍了Android Studio中主题样式的使用方法,本文通过实例代码给大家介绍的非常详细,对大家的学习或工作具有一定的参考借鉴价值,需要的朋友可以参考下
这篇文章主要介绍了AndroidManifest.xml中含盖的安全问题,本文通过实例代码给大家介绍的非常详细,对大家的学习或工作具有一定的参考借鉴价值,需要的朋友可以参考下
这篇文章主要介绍了AndroidStudio代码达到指定字符长度时自动换行实例,具有很好的参考价值,希望对大家有所帮助。一起跟随小编过来看看吧