Capturing Media With JavaScript

Oluwakeye John
August 08, 2020 · 1 min read


Have you ever wondered how a video chat app like zoom works? Well, it all boils down to capturing video and audio from a device.

Fortunately, JavaScript has an API called MediaDevices which can be used to access and use the media inputs of a device. The MediaDevices can be accessed under the window navigator object using navigation.mediaDevices.

Media input in this scenario can include camera stream, audio track, screen sharing service and others.


To start receiving the stream, we need to call the method MediaDevices.getUserMedia(). This method asks the user for permission to use the media input in question. The method resolves to the MediaStream object. If there is an error, the device doesn’t have the requested media, or the user denies the permission, then the promise is rejected.

const promise = navigator.mediaDevices.getUserMedia(constraints)

N.B. For security reasons, the getUserMedia method can only be used on a secure connection (secure includes HTTPS, localhost and file://). HTTP will not work, neither will an iframe.


The MediaDevices.getUserMedia takes an argument, called the constraints. The constraint describes the media type requested: video, audio or both. The constraint can also be used to specify more requirements on the returned media stream.

The following requests both video and audio.

const constraints = {
  video: true,

You can also specify additional requirements for each media type:

const constraints = {
  video: {
    width: 640,
    heigth: 480,
  audio: {
    noiseSuppression: true,

You can checkout a list of other constraints.

As earlier said, the getUserMedia() method returns a stream which you can decide to display on the device or transmit to another device for purposes like WebRTC(more to come on that later).

Here is a code snippet to access the MediaDevices API:

  .then(function (stream) {
    // process stream
  .catch(function (err) {
    // catch error

Basic example

The code snippet below gives an example of how to use the getUserMedia method. In this example, the stream is passed to the video srcObject property and displayed on the screen.

    <title>Capturing media with JavaScript </head>
    <video autoplay controls />

      const constraint = {
        video: true,
        audio: false,
      const video = document.querySelector("video")

        .then(stream => {
          video.srcObject = streams
        .catch(() => {

What next?

This post is just the tip of what you can do with the MediaDevices API. In a future post, I will:

  • Record the captured media using the MediaRecorder API
  • Stream the captured stream on another device via Web RTC

You have a suggestion, or you want to know more about something I mentioned above? Then feel free to contact me