Webinar: Accessible Video Players and UDL

By: Art Morgan
Instructional Design

Popular posts

business learning accessibility
A Guide to Accommodating Employees Who are Deaf A Guide to Accommodating Employees Who are Deaf
CaptionSync’s Kaltura MediaSpace integration
CaptionSync’s Kaltura MediaSpace Integration CaptionSync’s Kaltura MediaSpace Integration

Related posts

Webinar: Audio Description Webinar: Audio Description
Webinar: Using Cloud Storage for Accessible Video Webinar: Using Cloud Storage for Accessible Video

In this webinar we discuss the features of accessible video players, and how they can be used to implement the principles of universal design for learning (UDL).Annotated Video Transcript

>> Art Morgan: Hello, this is Art Morgan. I’ll be presenting today on the topic of accessible video players, and how video players tie in with the concept of universal design for learning. Kevin Erler will also be joining us for the Q&A session.

In this session we’ll review the characteristics of accessible video players and provide examples. We’ll use the CaptionSync Smart Player for most of our examples, but many video players follow these accessibility guidelines to varying degrees, and we’ll be happy to talk about the differences.

We’ll also learn about how accessible video fits into the universal design for learning framework, which is often referred to as UDL. And we’ll talk about how captions and audio description can enhance learning, sometimes in unanticipated ways.

Let’s start by taking a look at what the CaptionSync Smart Player is. At its core it’s an accessible video player, designed to meet WCAG 2.0 Level AA guidelines. It adds the interactive transcript functionality that you may already be used to from your YouTube account. It’s much more than just an interactive transcript player, however. It allows viewers to benefit from the caption and description data to easily navigate and search the video, look up terms in the video, and clip and share portions of the video, providing a more interactive learning experience.

Let’s talk briefly about how the Smart Player works. The video is streamed from its original source, which could be a URL to a video, such as a YouTube or Vimeo video, or it could be a video source that CaptionSync integrates to using URLs, such as Kaltura or Google Drive. We don’t make a local copy, which is why you need to provide us with a URL to the video to begin with. This is really important, because it allows you to caption and to play content in the Smart Player that you don’t own or otherwise have a copy of. The Smart Player then puts a frame around the streaming video and pulls the caption data directly from your CaptionSync account. It adds the captions below the video stream, and displays the interactive transcript to the side of the video. In addition, it uses the caption data to provide a number of other additional useful features that we’ll see in just a moment.

OK, let’s look at the features of the CaptionSync Smart Player, and how these features can make educational video more inclusive and more interactive. I have an example here, with the video pane on the left. Below the video is an area for traditional captions. We put the captions below the video so they don’t obscure anything in the video itself.

Below that we have the video controls. These controls can be controlled using the keyboard, which is important for people who can’t use a mouse, and they’re also high contrast, which is important for users with moderately low vision. Included in the controls are buttons for speeding up or slowing down the video, which many users find very useful. Some learners like to speed up the video to review certain content more quickly, and others like to slow it down.

The interactive transcript pane is on the right. This particular video is about exoplanets, and let’s say I’ve been assigned this video, and I’m not sure exactly what the term exoplanet means. I can set the player to “Show definition from dictionary,” select the word exoplanet, and I get a definition.

Now let’s say I know that one of the questions I need to answer is how astronomers can use the light spectrum of a star to determine the chemicals in the planet’s atmosphere. I can search for “spectrum,” and jump to an appropriate spot in the video to review that portion.

[ Video playing ]

>> By observing the light of a star during a transit, astronomers can find the fingerprint of the exoplanet’s atmosphere in the spectrum of the star. Each element creates distinctive dark lines, absorption lines in the spectrum. So these lines act as chemical fingerprints.

>> Now let’s imagine that you’re blind. The dialog in this video is quite informative, but it’s definitely not fully describing what’s being shown in the video. With the Smart Player and our audio description services, there’s a solution for that problem.

When audio description has been added to the video, we show the descriptions in the transcript pane with a light blue background. A blind user can either read the descriptions using her screen reader, or she can play the video and the player will insert audio descriptions created using text-to-speech. Here’s an example.

[ Text-to-speech voice ]

>> A spectrum line runs across the bottom of the screen with colors ranging from purple to blue to green, yellow, and orange. Black lines intersect the spectrum at various points and are identified as ozone, oxygen, water, methane, or CO2 markers. The intersecting black lines are of various widths.

[ Narrator’s voice ]

>> So these lines act as chemical fingerprints revealing the makeup of the atmosphere. Also, the stronger the line, the more of the corresponding element is present in the atmosphere.

>> Ok, so that’s audio description. Let me just review briefly a couple of the other control buttons below the video. The Preferences button, represented by a cog icon, allows you to hide the transcript pane, or configure it so that it appears underneath the video instead of to the side. And incidentally if you use the Smart Player on smaller screens it will automatically move the transcript pane below, in order to make both the video and the transcript more readable on tablets and mobile devices. Finally, the Clip and Share feature is useful if I’m a student and I want to share one portion of the video with a classmate or study group. I can select a specific portion of the video, select Generate URL, and then share that with my classmates.

So in summary, you can see that making video accessible isn’t just about adding captions and checking off a box that says your video is accessible. The Smart Player adds many benefits from a pedagogical standpoint. It’s consistent with universal design for learning principles in that it provides secondary representations of the content. It also provides additional ways for learners to engage with the content in non-linear ways that you don’t get with traditional video.

One other powerful feature of the Smart Player is the ability to embed the player right in to your own webpage. Once you have a Smart Player link, you can use the Embed button to generate an embed code, which is just a simple snippet of iframe code that you can drop into your own webpage and make the Smart Player appear right on your own page. This feature is useful for embedding the Smart Player right into your LMS pages. That’s a topic we cover in more detail in a separate webinar session, but I’m showing here an image of the Smart Player embedded in a D2L Brightspace page.

One other key feature of the Smart Player is the ability to enable you to caption other people’s YouTube videos. If you’ve ever had the need to present somebody else’s YouTube video and make it accessible, you’ll know that this can be quite a challenge, but the Smart Player makes it easy. We also cover this in more detail in a separate webinar session.

Finally, let’s review some of the ways captions and descriptions can enhance learning. In a previous webinar I mentioned research by Dr. Bryan Dallas that showed that students who viewed a 15 minute video with captions performed better on an assessment test than those who viewed the same video without captions. This is consistent with the UDL principle that says that providing additional representations of the content, in this case a written representation of the audio content, can help many learners understand the content better. Audio description is less frequently thought of as a secondary representation — most people think of it as an accommodation — but you can imagine how the same principles might apply. In other words, some people, even if they are not blind, might benefit by being able to read a description of what they are seeing, or hear a description. The descriptions can augment people’s understanding of the visual content.

I’m showing on this slide an image of the brain with the text, Recognition Networks: The What of Learning. This image is from CAST.org, which is an excellent resource for more information about universal design for learning.

That’s it for this session. Our resource links include a link to information about the Able Player, which is an open source accessible video player, as well as detailed info about the Smart Player on our Support Center site. Our blog has links to past webinars and other video accessibility articles. If you have any further questions, please let us know.