Stream recording¶
Overview¶
A media stream captured by WCS can be recorded during publishing.
Supported protocols¶
- WebRTC
- RTMP
- RTSP
- MPEG-TS
Recording formats¶
- MP4 for H.264 + AAC codecs
- WebM for VP8 + Opus codecs
- WebM for H264 + opus codecs (since WCS build 2.0.2001)
- TS for H.264 + ADTS
- MKV (since WCS build 5.2.1190)
Quick manual on testing¶
-
For the test we use the demo server
demo.flashphoner.com
and the Stream Recording web application
-
Click the
Start
button. Capturing and publishing of the stream starts
-
Click the
Stop
button. Broadcasting stops, and a link to play and download the recorded fragment appears
Configuration¶
Server side¶
Turning stream recording on and off¶
By default, stream recording is turned on.
To turn recording off, set the following parameter in flashphoner.properties file:
Parameter
turns on recording for all the streams published with RTMP encoder or republished from another RTMP server.
Parameter
turns on recording for all the streams captured from RTSP IP cameras.
Parameter
turns on recording for all the mixers output streams.
MP4, WebM, MKV containers support¶
By default, H264 streams are recorded to MP4 container and VP8 streams to WebM one. Since WCS build 5.2.1190, MKV contaner is supported which is a swiss army knife at codecs point but cannot be played directly in browser.
The following parameter is used to set recording containers
MKV recording can be enabled as follows
One of the codecs published may be recorded to MKV container (VP8 for example)
Since build 2.0.2001, WebM container is supported for WebRTC H264+opus streams recording
Attention
Audio only streams are always recorded to MP4 container (using AAC codec)
Recording to MPEG-TS container¶
Since build 5.2.610, H264 streams can also be recorded to TS container using the following parameter
or since build 5.2.1190
The record_h264_to_ts
parameter has a priority above record_formats
, i.e.
makes H264 streams to be recorded to MPEG-TS.
Known limits¶
-
VLC before 3.0.8 can not play TS recordings.
-
TS recordings can not be rewound while playing in VLC.
Forming the name of the stream record file¶
By default, the file name is formed by template specified with stream_record_policy_template
parameter:
The following tags can be used in template:
Tag | Description | Maximum size |
---|---|---|
{streamName} |
Stream name | Unlimited |
{duration} |
File duration, for MP4 recordings only | Unlimited |
{startTime} |
Recording start time | 20 characters |
{endTime} |
Recording end time | 20 characters |
{startTimeMillis} |
Recording start time based on server clock | 20 characters |
{endTimeMillis} |
Recording end time based on server clock | 20 characters |
{sessionId} |
Session ID (BASE64 encoded) | 60 characters |
{mediaSessionId} |
Media session ID | 36 characters |
{login} |
Login | 32 characters |
{audioCodec} |
Audiocodec | 4 characters |
{videoCodec} |
Videocodec | 4 characters |
For example,
means that the file name will match the stream name. So, the stream published with ffmpeg
ffmpeg -re -i BigBuckBunny.mp4 -preset ultrafast -acodec aac -vcodec h264 -strict -2 -f flv rtmp://test1.flashphoner.com:1935/live/stream_ffmpeg
will be written to file stream_ffmpeg.mp4
.
File extension is set depending on stream parameters and container used: mp4
, webm
, mkv
or ts
.
When the file name matches the stream name, it may contain characters that are not allowed in file names, slash /
for example. In that case, the file name should be encoded using the parameter
Then, the file name will be encoded with a hexadecimal number. The parameter
will encode the file name with BASE64 encoding.
Another way to escape invalid characters is to remove them using exclude_record_name_characters
parameter. By default
For example, to remove colons, commas, periods and slashes set
Record files rotation¶
Stream records can be splitted to parts of a given duration using record_rotation
parameter. For example, the setting
specifies a fragment duration as 20 seconds and setting
defines a fragment maximum volume as 10 megabytes.
If recording file name template contains {startTime}
tag, recording fragment start timestamp will be inserted into file name. If template contains {endTime}
tag, recording fragment end timestamp will be inserted into file name. For example, if the following settings are used
the test
stream recording fragments will be named as follows
test-1553577643961-1553577663988_1.mp4
test-1553577663989-1553577683997_2.mp4
test-1553577683997-1553577701626_3.mp4
...
Recording fragments are numbered continuously from 1. For every new mediasession (even if stream is published with the same name) indexing starts again, i.e. old fragments will be overwritten if static part of file name template is not unic (stream name only for example).
Indexing can be disabled if necessary with the following parameter
In this case recording fragments are not numbered and will be named exactly as template sets. If template does not provide unic names, old fragments may be overwritten.
Recording start time, end time and duration time calculation¶
Since build 5.2.458, recording start time, end time and duration time are calculated by stream frames time stamps. note that RTMP stream timestamps are always starting from 0, but WebRTC publisher usually sets full time stamp by its own clock for ever frame.
Since build 5.2.635 it is possible to use start time and end time based on server clock
In general, stream time stamps are different from server clock time stamps.
To calculate recording times more precisely, audio data should be buffered to keep synchronization. To do this, the following parameter is added
By default, audio buffer size is set to 100 packets.
Record files handling script¶
The on_record_hook_script
setting points to the shell script that is invoked when stream recording finishes.
The script is placed to the /usr/local/FlashphonerWebCallServer/bin
folder by default:
but it can be placed to any folder with any name, for example:
This script can be used to copy or move the stream record from the /usr/local/FlashphonerWebCallServer/records
folder to another location after recording completes.
Example:
STREAM_NAME=$1
SRC_FILE=$2
SRC_DIR="/usr/local/FlashphonerWebCallServer/records/"
REPLACE_STR="/var/www/html/stream_records/$STREAM_NAME-"
DST_FILE="${SRC_FILE/$SRC_DIR/$REPLACE_STR}"
cp $SRC_FILE $DST_FILE
Here
$1
- stream name$2
- absolute path and file name of the stream record- when stream recording ends, the record file is copied to
/var/www/html/stream_records/
It is necessary to take into account the length of the absolute file name (including folder path) that will be formed when copying record file. If the absolute name of the target file exceeds 255 characters limit, copy command will fail with error, so the handling script will not work as expected.
Since build 5.2.801, WCS is running from flashphoner
user for security reasons. Therefore, if recording files are moved to other folder, it is necessary to allow writing to the folder. For example, if files are copied to /opt/media
folder
STREAM_NAME=$1
FILE_NAME=$2
echo $STREAM_NAME:$FILE_NAME >> /usr/local/FlashphonerWebCallServer/logs/record.log
cp $FILE_NAME /opt/media
writing permissions must be granted to this folder
Directory for saving recorded files¶
By default, stream records are saved to the folder /usr/local/FlashphonerWebCallServer/records
. Since build 5.2.687, the folder for saving records can be changed using the following parameter
If this parameter defines a non-default folder, files downloading in Stream Recording example may fail. In this case, it is recommended to deploy a separate web server for recording files downloading from a custom folder.
Since build 5.2.801, WCS is running from flashphoner
user for security reasons. Therefore, when this parameter is changing, write permissions must be granted to the folder as described above.
Adjusting record audio sample rate¶
By default, audio track is recorded with sample rate 44.1 kHz. This value can be changed using the following parameter if necessary
In this case, record audio sample rate will be set to 48 kHz.
moov
atom placement in recording metadata¶
To play a recording file while downloading (progressive downloading), moov
atom must be placed before mdat
atom in recording metadata. To do this, the following default setting is added in latest builds
The place for moov
atom can be reserved when recording file is created to optimize disk operations while saving the file. This feature can be enabled with the following parameter
A space size to reserve should be set in kilobytes with the following parameter
By default, 2048 kilobytes will be reserved. Therefore, a recording file size will not be less than reserved space if moov
atom reservation is enabled, this must be taken into account while setting up record rotation by size.
Audio track bitrate configuration for FDK AAC codec¶
Since 5.2.428, the ability is added to set audio track bitrate mode for recording with FDK AAC codec. By default, mode 5 is set (VBR 112 kbps). This value can be changed with the following parameter
The following bitrate modes are possible:
- 0 - CBR
- 1-5 - VBR
Note that recording files playback by file timestamps using nginx module ngx_http_mp4_module is possible for VBR recordings only.
Audio channels configuration¶
Since build 5.2.610, it is possible to set audio channels count for recordings using the following parameter
By default, audio channels count set to 2 (stereo). To record stream with monophonic audio, the following should be set
Recording perfomance tuning under high load¶
Recording files writing to disk may take a much time when a number of published streams are recorded simultaneously. To speed up writing, it is possible since build 5.2.639 to set CPU threads number to work with stream recording using the following parameter
By default, 4 threads are used to work with stream recording. This number can be increased if necessary.
Warning
It is not recommended to set recording threads number more than CPU cores available on the server
VP8 streams recording to webm container¶
Since build 5.2.905 java implementatiion is used to record VP8 streams to webm container
A cluster duration (in milliseconds) and size (in bytes) limits settings are available for this implementation. A stream data will be flushed to a recording file when one of the limits is reached
If some issues occur, the native code recorder implementation based on ffmpeg may be enabled
A separate folder for temporary files¶
While recording a stream, media data are written to a temporary file, then this file is copying to a recording file named according to template. Since build 5.2.963, it is possible to set a separate folder for temporary files using the following parameter
This allows, for example, to place temporary files to RAM drive to speed up recording process.
The folder to place temporary files must be writable by WCS process, so permissions must be set to flashphoner
user. For example, if the temporary files folder is set to
then permissions must be set to flashphoner
By default, all the temporary files are placed to /usr/local/FlashphonerWebCallServer/records
folder.
Minimal available disk space checking¶
Since build 5.2.1209 a minimal available disk space is checked when recording a stream. If the space available is less then a limit, the recording will stop, or will not be started. In this case, the following message will be written to server log
The space limit is set by the following parameter (1 G by default)
It is possible to set the limit value in gigabytes ( g
suffix) or in megabytes ( m
suffix), for example
Attention
If the recording was stopped due to space limit, a post processing script still will be launched for the recording file
Recording stopping on errors¶
Since build 5.2.1236 the parameters were added to define an errors handling while recording a stream. By default, stream recording will be stopped if 3 errors occur during 60 minutes:
The frame which raise an exception and subsequent frames will not be recorded until keyframe is received and parsed successfully.
Updating the MP4 container atoms with actual recording data¶
Since build 5.2.1990 it is possible to update MP4 atoms data on the fly
In this case MP4 atoms data are updating periodically according to actual recording media data. An atoms update interval is set in seconds
This feature allows to play the recording file even if the recording was interrupted due to some error. The media data in such MP4 file will be available to play according to MP4 atoms.
Recording file rotation by MP4 header filling¶
Since build 5.2.2012 a recording MP4 file will automatically be rotated when a space is reserved for MP4 header in file and MOOV atom becomes filled. For example, if a 1920x1080 stream is publishing with bitrate 3 Mbps and 30 fps, and the stream is recording with the following settings
then one recording part should be near 22 minutes long (near 510 Mb in size).
Updating MP4 header without space reservation¶
It is necessary to reserve a huge space for MP4 file header for a long streams recording. For example, moire than 5 M should be reserved to record a 1080p stream for one hour, and a much space will be left empty for a shorter recordings. To resolve this, since build 5.2.2020 MP4 hedaer actualization flow was changed. With the following parameter
and other settings left by default, the MP4 header will be written to a separate file. Thus, the recording consist of two files:
-rw-r--r-- 1 flashphoner flashphoner 45293 Jul 15 09:58 test-f6a0c740-4255-11ef-af8d-e781f7a0d3ed.mp4.header
-rw-r--r-- 1 flashphoner flashphoner 924777 Jul 15 09:58 test-f6a0c740-4255-11ef-af8d-e781f7a0d3ed.mp4.tmp
A header data are writtent to .header
file and a mediadata are written to .tmp
file. The files are concatenated when recording is finished
-rw-r--r-- 1 flashphoner flashphoner 1930149 Jul 15 09:58 test-f6a0c740-4255-11ef-af8d-e781f7a0d3ed.mp4
If the recording was occasionally stopped by some failure, the header file and the data file remain on disk. In this case the files should be concatenated manually to play the data recorded
cat test-f6a0c740-4255-11ef-af8d-e781f7a0d3ed.mp4.header test-f6a0c740-4255-11ef-af8d-e781f7a0d3ed.mp4.tmp > test-f6a0c740-4255-11ef-af8d-e781f7a0d3ed.mp4
Client side¶
If stream recording is enabled on the server, whether the stream is recorded or not is determined by the value of record
parameter passed into the createStream function in the script of the publisher client:
true
- the stream published by this client is recorded;false
- the stream is not recorded.
For instance, the script of the Stream Recording application recording.html, recording.js, contains the following code:
function publishStream(session) {
var streamName = $('#url').val().split('/')[3];
session.createStream({
name: streamName,
display: localVideo,
record: true,
receiveVideo: false,
receiveAudio: false
...
}).publish();
}
Stream recording on demand¶
Sometimes, it is necessary to record the stream that already exists on server, mixer output stream for example. This can be done with REST API. Note that only streams in PUBLISHING
state can be recorded.
REST query must be HTTP/HTTPS POST query like this:
- HTTP:
http://streaming.flashphoner.com:8081/rest-api/recorder/startup
- HTTPS:
https://streaming.flashphoner.com:8444/rest-api/recorder/startup
Where:
streaming.flashphoner.com
is WCS server address8081
is a standard WCS REST / HTTP port8444
is a standard WCS REST / HTTPS portrest-api
is mandatory prefix/recorder/startup
is REST method
REST methods and responses¶
/stream/startRecording, /recorder/startup¶
Start stream recording in specified mediasession
Request example¶
POST /rest-api/recorder/startup HTTP/1.1
Host: localhost:8081
Content-Type: application/json
{
"mediaSessionId": "5a072377-73c1-4caf-abd3",
"config": {
"fileTemplate": "{streamName}-{startTime}-{endTime}",
"rotation": "20M"
}
}
Response example¶
Return codes¶
Code | Reason |
---|---|
200 | OK |
404 | Not found |
500 | Internal error |
/stream/stopRecording, /recorder/terminate¶
Stop stream recording in specified mediasession
Request example¶
POST /rest-api/recorder/terminate HTTP/1.1
Host: localhost:8081
Content-Type: application/json
{
"mediaSessionId": "5a072377-73c1-4caf-abd3"
}
Return codes¶
Code | Reason |
---|---|
200 | OK |
404 | Not found |
500 | Internal error |
/recorder/find_all¶
Find session recordings
Request example¶
Response example¶
HTTP/1.1 200 OK
Access-Control-Allow-Origin: *
Content-Type: application/json
[
{
"fileName": "9c3e-test-1563776083752-{endTime}.mp4",
"mediaSessionId": "5a072377-73c1-4caf-abd3"
}
]
Return codes¶
Code | Reason |
---|---|
200 | OK |
404 | Not found |
500 | Internal error |
Parameters¶
Parameter | Description | Example |
---|---|---|
mediaSessionId | Media session identificator | 5a072377-73c1-4caf-abd3 |
config | Record settings that redefine server settings | |
fileTemplate | Recording file name template | {streamName}-{startTime}-{endTime} |
rotation | Enables/disables rotation and duration/volume of fragments | 20M |
Recording on demand works as follows:
- When REST API
/recorder/startup
query is called, current recording for the stream will be stopped. - New recording starts with settings passed in REST query.
- If some setting is not defined in REST query, the server setting will be applied.
For example, if recording should have exact file name with rotation disabled, the following query should be passed:
/stream/startRecording
{
"mediaSessionId":"1234567890abcdefgh",
"config": {
"fileTemplate": "test",
"rotation": "disabled"
}
}
REST query /recorder/find_all
returns session recordings list. The list shows both recordings on demand started via REST API and recordings
initiated with WebSDK:
[
{
"fileName": "003f-1563776713987-{endTime}.mp4",
"mediaSessionId": "5af9c820-ac49-11e9-9f06-693cb47c4042"
},
{
"fileName": "stream-57882100-ac49-11e9-afdd-6752f5be57a9-jtdvnittjkrd8rsc3dnfbger2o.mp4",
"mediaSessionId": "57882100-ac49-11e9-afdd-6752f5be57a9"
}
]
How to get recording file name¶
There are the following ways to get recording name, for example, to download it:
-
Record files handling script on server receives the recording file name right after it recorded
-
If it is necessary to know recording name in browser, file name template should be formed from parameters which can be obtained by REST API, for example
- When WebSDK is used, recording name can be obtained with
getRecordInfo()
function
...
}).on(STREAM_STATUS.UNPUBLISHED, function (stream) {
setStatus(stream.status());
showDownloadLink(stream.getRecordInfo());
onStopped();
});
...
Note that STREAM_STATUS.UNPUBLISHED
may be received in considerable time after stopping a stream if recording file is big. Since build 5.2.673 this time may be limited using the following parameter (15 seconds by default)
How to play or download recording file¶
The recording file is available via WCS internal web server using the following link
Where
test.flashphoner.com
- WCS server URLstream.mp4
- recording file name
By default, WCS returns HTTP header
in this case, browser tries to play the file. This behaviour is enabled with the following parameter
To make a browser to download the recording file without playing it, the following parameter should be set
Downloading and playing a certain part of a recording file¶
Since build 5.2.894 it is possible to download and play a certain part of a recording file. To do this, request the file with start time and end time in seconds
Only start time or only end time can be used.
The fragments requested are stored to the same folder where recording files are placed. Timestamps will be added to the fragment name, for example
Such files are not removed after downloading: if the same fragment is requested again, server will send the existing file.
Since build 5.2.899 partial download and playback is supported for audio only recordings too.
Fragments folder configuration¶
By default, fragments are written to the folder /usr/local/FlashphonerWebCallServer/records
(the same folder as recording files).
Since build 5.2.957 it is possible to set a separate folder to store fragments using the following parameter
Known limits¶
-
A partial download and playback is supported for MP4 container only. If webm file is requested, it will be always fully downloaded and played.
-
Playback can start slightly earlier if start time is set, depending on closest key frame in the recording file.
Multiple stream recording to one file with subsequent mixing¶
Since build 5.2.1012 it is possible to record multiple streams to one file. Then streams can be extracted from this file and mixed by a special tool. Multiple streams can be recorded to MP4 container or, since build 5.2.1440, to MKV container. This feature is intended, for example, to record a video conference. In this case, unlike to MCU mixer, stream mixing works while recording file is post-processed, this allows to aquire a less of server resources during the conference itself.
Multiple stream recording is managed by REST API.
Codecs support¶
MP4 container:
- H264
- AAC
MKV container:
- H264
- VP8
- Opus
- AAC
- PCMA
- PCMU
- G722
Multirecorder REST API¶
REST query must be HTTP/HTTPS POST query like this:
- HTTP:
http://streaming.flashphoner.com:8081/rest-api/multipleRecorder/startup
- HTTPS:
https://streaming.flashphoner.com:8444/rest-api/multipleRecorder/startup
Where:
streaming.flashphoner.com
is WCS server address8081
is a standard WCS REST / HTTP port8444
is a standard WCS REST / HTTPS portrest-api
is mandatory prefix/multipleRecorder/startup
is REST method
/multipleRecorder/startup¶
Launch multiple streams recorder
Request example¶
POST /rest-api/multipleRecorder/startup HTTP/1.1
Host: localhost:8081
Content-Type: application/json
{
"uri": "multi-recorder://test-record"
}
Response example¶
Return codes¶
Code | Reason |
---|---|
200 | OK |
409 | Conflict |
500 | Internal error |
/multipleRecorder/add¶
Add the stream with certain mediaSessionId to the recorder
Request example¶
POST /rest-api/multipleRecorder/add HTTP/1.1
Host: localhost:8081
Content-Type: application/json
{
"uri": "multi-recorder://test-record",
"mediaSessionId": "866a9910-fbfe-11eb-aae4-6f99b0c80a3a"
}
Response example¶
Return codes¶
Code | Reason |
---|---|
200 | OK |
404 | Not found |
409 | Conflict |
500 | Internal error |
/multipleRecorder/find_all¶
Find all recorders
Request example¶
POST /rest-api/multipleRecorder/find_all HTTP/1.1
Host: localhost:8081
Content-Type: application/json
Response example¶
HTTP/1.1 200 OK
Access-Control-Allow-Origin: *
Content-Type: application/json
[
{
"mediaSessionsId": [
"866a9910-fbfe-11eb-aae4-6f99b0c80a3a",
"9f1e2530-fbfe-11eb-9ec1-77172ac14d86",
"a970d0a0-fbfe-11eb-8fcc-912807bab442"
],
"uri": "multi-recorder://test-record",
"fileName": "multi-recorder___test-record.mp4"
}
]
Return codes¶
Code | Reason |
---|---|
200 | OK |
404 | Not found |
500 | Internal error |
/multipleRecorder/remove¶
Remove stream with certain mediaSessionId from recorder
Request example¶
POST /rest-api/multipleRecorder/remove HTTP/1.1
Host: localhost:8081
Content-Type: application/json
{
"uri": "multi-recorder://test-record",
"mediaSessionId": "866a9910-fbfe-11eb-aae4-6f99b0c80a3a"
}
Response example¶
Return codes¶
Code | Reason |
---|---|
200 | OK |
404 | Not found |
500 | Internal error |
/multipleRecorder/terminate¶
Stop multiple streams recorder
Request example¶
POST /rest-api/multipleRecorder/terminate HTTP/1.1
Host: localhost:8081
Content-Type: application/json
{
"uri": "multi-recorder://test-record"
}
Response example¶
Return codes¶
Code | Reason |
---|---|
200 | OK |
404 | Not found |
500 | Internal error |
Parameters¶
Parameter | Description | Example |
---|---|---|
uri | Recorder URI | multi-recorder://test-record |
mediaSessionId | Stream mediasession Id | 866a9910-fbfe-11eb-aae4-6f99b0c80a3a |
filename | Recording file name | multi-recorder___test-record.mp4 |
Recording file name¶
Multiple streams recording file name is formed by template by the following rules:
-
{streamName}
parameter is set according to recorder URI, replacing all the characters not allowed to use in file system to underline. -
{startTime}
,{endTime}
parameters cannot be set because they depend on stream timestamps, and we have a multiple streams with a different timestamps simultaneously. So it is recommended to use{startTimeMillis}
,{endTimeMillis}
parameters to add servers clock timestamps to file name.
For example, with the following template
the file name for recorder with URI
will look as follows:
Here {startTime}
, {endTime}
are replaced by -1
.
Multiple streams recording folder¶
By default, multiple streams recording files are stored in /usr/local/FlashphonerWebCallServer/records
folder. Since build 5.2.1088, multiple streams recording folder can be set using the following parameter
This folder should be writalble, For example, if the folder is set to
the folder access privilegies sholud be set as
Multiple streams mixing tool¶
Only one stream can be played by default from multiple streams recording file. The streams should be mixed to view them all. The OfflineMixer tool should be used to mix, launching as follows:
cd /usr/local/FlashphonerWebCallServer/tools
./offline_mixer_tool.sh multi-recorder___test-record--1-1628821032180--1-1628821151750.mp4
Offline mixer settings should be defined in /usr/local/FlashphonerWebCallServer/conf/offline_mixer.json
file. By default, the following settings are used:
A mixed file is placed to the same folder as original one, with _mixed
suffix addition, for example
A picture sample from the mixed file
Getting tracks information from multiple recording file¶
Since build 5.2.1049 tracks information from multiple recording file may be extracted using multiple stream mixing tool. To do this, the tool should be launched as follows:
In this case, the tool will print tracks information in JSON form. See two participants room recording information example below:
Two participants track information example
[
{
"durationInMS": "37282",
"trackEdits": [
{
"endInMs": "14",
"type": "pause",
"startInMs": "0"
},
{
"endInMs": "37282",
"type": "media",
"startInMs": "14"
}
],
"channels": "2",
"trackType": "AUDIO",
"trackId": "1",
"timescale": "44100",
"streamName": "room-09f012-user1-e6ff",
"trackCodec": "mp4a",
"sampleRate": "44100",
"mediaSessionId": "e6ff54e0-1c2a-11ec-90e8-79a2a32f3d9d"
},
{
"durationInMS": "37336",
"trackEdits": [
{
"endInMs": "37336",
"type": "media",
"startInMs": "0"
}
],
"trackType": "VIDEO",
"trackId": "0",
"width": "320",
"timescale": "90000",
"streamName": "room-09f012-user1-e6ff",
"trackCodec": "avc1",
"mediaSessionId": "e6ff54e0-1c2a-11ec-90e8-79a2a32f3d9d",
"height": "240"
},
{
"durationInMS": "39274",
"trackEdits": [
{
"endInMs": "100",
"type": "pause",
"startInMs": "0"
},
{
"endInMs": "21534",
"type": "pause",
"startInMs": "100"
},
{
"endInMs": "39274",
"type": "media",
"startInMs": "21534"
}
],
"channels": "2",
"trackType": "AUDIO",
"trackId": "3",
"timescale": "44100",
"streamName": "room-09f012-user2-f746",
"trackCodec": "mp4a",
"sampleRate": "44100",
"mediaSessionId": "f74633a1-1c2a-11ec-bba5-af8cf43275a8"
},
{
"durationInMS": "39303",
"trackEdits": [
{
"endInMs": "21434",
"type": "pause",
"startInMs": "0"
},
{
"endInMs": "39303",
"type": "media",
"startInMs": "21434"
}
],
"trackType": "VIDEO",
"trackId": "2",
"width": "320",
"timescale": "90000",
"streamName": "room-09f012-user2-f746",
"trackCodec": "avc1",
"mediaSessionId": "f74633a1-1c2a-11ec-bba5-af8cf43275a8",
"height": "240"
}
]
Where:
durationInMS
- track duration in millisecondstrackType
- track type:AUDIO
orVIDEO
trackId
- track IdstreamName
- stream name containing this trackmediaSessionId
- stream media session Idtimescale
- track samples per second quantitytrackCodec
- track codecwidth
,height
- video track picture size by first key framechannels
- audio track channels countsampleRate
- audio track sample rate (usually equal totimescale
parameter)trackEdits
- track timeline description
Track timeline is described as a set of segments according to MP4 edit lists
atom content, with the following parameters:
startInMs
- segment start time in milliseconds relative to file beginningendInMs
- segment end time in milliseconds relative to file beginningtype
- segment type: mediadata (media
) or pause (pause
)
Any single track may be extracted from multiple recording file using ffmpeg or other MP4 editing tool.
Note that if a stream with the same name was added to multiple recorder, then removed from recorder and then added again, this stream will be repersented in the file as different tracks with subsequent track Ids, for example:
Re-added stream track information example
[
{
"durationInMS": "78978",
"trackEdits": [
{
"endInMs": "63050",
"type": "pause",
"startInMs": "0"
},
{
"endInMs": "78978",
"type": "media",
"startInMs": "63050"
}
],
"channels": "2",
"trackType": "AUDIO",
"trackId": "3",
"timescale": "44100",
"streamName": "test",
"trackCodec": "mp4a",
"sampleRate": "44100",
"mediaSessionId": "fbbf5b50-20ee-11ec-bf06-ef6ec6048b2c"
},
{
"durationInMS": "39708",
"trackEdits": [
{
"endInMs": "23150",
"type": "media",
"startInMs": "0"
}
],
"channels": "2",
"trackType": "AUDIO",
"trackId": "1",
"timescale": "44100",
"streamName": "test",
"trackCodec": "mp4a",
"sampleRate": "44100",
"mediaSessionId": "c7bc1460-20ee-11ec-bf06-ef6ec6048b2c"
},
{
"durationInMS": "39791",
"trackEdits": [
{
"endInMs": "23233",
"type": "media",
"startInMs": "0"
}
],
"trackType": "VIDEO",
"trackId": "0",
"width": "640",
"timescale": "90000",
"streamName": "test",
"trackCodec": "avc1",
"mediaSessionId": "c7bc1460-20ee-11ec-bf06-ef6ec6048b2c",
"height": "360"
},
{
"durationInMS": "63050",
"trackEdits": [
{
"endInMs": "39791",
"type": "pause",
"startInMs": "0"
},
{
"endInMs": "50191",
"type": "media",
"startInMs": "39791"
}
],
"trackType": "VIDEO",
"trackId": "2",
"width": "640",
"timescale": "90000",
"streamName": "test",
"trackCodec": "avc1",
"mediaSessionId": "ed3ebda0-20ee-11ec-bf06-ef6ec6048b2c",
"height": "360"
}
]
Pulling a separate streams from MKV container¶
Since build 5.2.1440 a separate streams can be pulled from MKV container using multiple stream mixing tool:
In this case, MKV files will be created per each stream in multirecording:
multi-recorder___test-record_fbbf5b50-20ee-11ec-bf06-ef6ec6048b2c.mkv
multi-recorder___test-record_c7bc1460-20ee-11ec-bf06-ef6ec6048b2c.mkv
multi-recorder___test-record_ed3ebda0-20ee-11ec-bf06-ef6ec6048b2c.mkv
If a stream was removed from multirecorder and when added again, or was added a much later then other streams, the gaps will be filled by default to equalize all the pulled streams. This can be disabled if necessary
Multiple recording file handling script¶
When multiple stream recording is finished, a specila handling script is automatically launched as set by the following parameter
By default, script will launch offline_mixer_tool.sh
, passing multiple streams recording file name to it.
Since build 5.2.1023, on_multiple_record_hook.sh
script writes only file mixing result to log file /usr/local/FlashphonerWebCallServer/logs/multi-record.log
, to decrease hard disk I/O load while offline mixing tool is working. A detailed logging may be enabled if necessary by setting the following variable in the script
Multi-threaded encoding while mixing multiple stream recordings¶
Since build 5.2.1089 multi-threaded encoding can be enabled while mixing multiple stream
recordings. To enable this feature, add the following parameter to /usr/local/FlashphonerWebCallServer/conf/offline_mixer.json
file
Multiple stream recordings are mixed a twice faster when multithreaded encoding is enabled comparing to singe-threaded one.
Threads count to use for multi-threaded encoding¶
Since build 5.2.1523 threads count for multi-threaded encoding can be set up. By default, threads count is equal to a half of CPUs available. For example, on 12 vCPU server 6 threads will be used
If mixing takes a long time, threads count may be increased, but it is not recommended to set it more than CPUs count, which can be detected by the following command
Stream name displaying in mixed multiple record¶
By default, every stream name is displayed in mixed multiple record. This may be disabled if necessary by the following setting in /usr/local/FlashphonerWebCallServer/conf/offline_mixer.json
file
When RoomApi conference streams are recorded, stream name includes a room name and a participant stream id, for example room-1882a6-bob-037c
. Since build 5.2.1642 room name may be excluded by the following settings
{
...,
"mixerDisplayStreamName": true,
"mixerTextDisplayRoom": false,
"labelReplaceRegex": "\\w+-\\w+-([^\\-]+)-\\w+",
"labelReplaceWith":""
}
Where:
labelReplaceRegex
- a regular expression to find items to be replaced in a stream namelabelReplaceWith
- string to replace items found, empty string will exclude the items
In this case, for the example above, only participant name bob
will be displayed.
Characters decoding in stream name¶
Since build 5.2.1751 a characters encoded at client side by encodeURIComponent() may be decoded during mixing a multiple record
In this case a decoded characters available in the font used will be displayed, or a similar characters.
Multiple recording data callback¶
Since build 5.2.1123 it is possible to send PORT query to a certain URL after multiple recording and mixing is finished. This allows to notify backend about file name to which chat room recording is mixed.
Callback URL should be set in /usr/local/FlashphonerWebCallServer/conf/offline_mixer.json
file
A recording data to send are passed via /usr/local/FlashphonerWebCallServer/bin/on_multiple_record_hook.sh
script when launching offline_mixer_tool.sh. Therefore, if WCS build
5.2.1123 is installed over a previous build, or if custom on_multiple_record_hook.sh
is used, the script should be modified as follows:
on_multiple_record_hook.sh
# This script copies a recorded stream to client/records
FILE_NAME=$1
CREATION_MODIFICATION_TIME=$2
DURATION_IN_MS=$3
RECORDER_URI=$4
WCS_HOME=/usr/local/FlashphonerWebCallServer
LOG_FILE=$WCS_HOME/logs/multi-record.log
MIXER_TOOL=$WCS_HOME/tools/offline_mixer_tool.sh
# Set LOGGER_ENABLED to true to enable mixing debug logging
LOGGER_ENABLED=false
echo "[$(date '+%Y-%m-%d %H:%M:%S')] Start mixing multiple recording file $FILE_NAME" >> $LOG_FILE
if $LOGGER_ENABLED; then
bash $MIXER_TOOL $FILE_NAME $CREATION_MODIFICATION_TIME $DURATION_IN_MS $RECORDER_URI >> $LOG_FILE 2>&1
else
bash $MIXER_TOOL $FILE_NAME $CREATION_MODIFICATION_TIME $DURATION_IN_MS $RECORDER_URI > /dev/null 2>&1
fi
CODE=$?
if [ "$CODE" -ne "0" ]; then
if [ "$CODE" -eq "64" ]; then
echo "ERROR: File to mix not found" >> $LOG_FILE
elif [ "$CODE" -eq "65" ]; then
echo "ERROR: Offline mixer config not found" >> $LOG_FILE
else
echo "ERROR: Offline mixer tool error code: $CODE" >> $LOG_FILE
fi
exit $CODE
fi
echo "[$(date '+%Y-%m-%d %H:%M:%S')] Multiple recording file $FILE_NAME is mixed successfully" >> $LOG_FILE
exit 0
POST query contains a following JSON data:
POST /multiRecorderCallback HTTP/1.1
Content-Type: application/json
Content-Length: 463
Host: localhost
Connection: Keep-Alive
User-Agent: Apache-HttpClient/4.3.5 (java 1.5)
Accept-Encoding: gzip,deflate
{
"multiRecorderCreationModificationTime":3724973476,
"multiRecorderDurationInMS":44061,
"multiRecorderFilePath":"/usr/local/FlashphonerWebCallServer/multirecords/stream-32c7edd7-37bf-4bf2-a58d-955679c5287e-mockLogin.mp4",
"recorderUri":"multi-recorder://room-bace1f",
"mixerParams":
[
{
"path":"/usr/local/FlashphonerWebCallServer/multirecords/stream-32c7edd7-37bf-4bf2-a58d-955679c5287e-mockLogin_mixed.mp4",
"durationInMs":44000,
"creationModificationTime":3724973524
}
]
}
Multiple recording file parameters:
multiRecorderCreationModificationTime
- multiple recording file creation date and timemultiRecorderDurationInMS
- multiple recording file duration in millisecondsmultiRecorderFilePath
- multiple recording file pathrecorderUri
- multiple recording identifier, contains room name if RoomApi is used
Mixed file parameters:
path
- mixed file pathdurationInMs
- mixed file duration in millisecondscreationModificationTime
- mixed file creation date and time
Minimal available disk space checking for multiple recordings mixing¶
Since build 5.2.1317 minimal available disk space checking is enabled for multiple recordings mixing. The mixing will not start or will be stopped when disk space becomes less then defined threshold value. The threshold is set by the following parameter in /usr/local/FlashphonerWebCallServer/conf/offline_mixer.json
file
By default, disk space threshold is set to 1 Gb (same as for single stream recordings). If the treshold is reached while mixing is already working, the mixing will be stopped with keeping data already written to destination file. The file can be correctly played after that.
Known issues¶
1. Beware of maximum file name system limit¶
Maximum length of file name in all actual Linux file systems is limited to 255 characters. When record file is created, its name will be trimmed to this limit including extension and part number if rotation is enabled.
2. Creation time will be set to MP4 recording metadata only¶
File creation time will be set to MP4 recording metadata only, but not Webm or MKV.
3. All the target folders to copy or move recordings should be writable¶
Symptoms
Record hook script does not perform any operation on record files
Solution
Grant write access to a target folder. For example, if script does
allow others
group to write to
4. Two channels audio recording increases server CPU load¶
CPU load may grow and simultaneous streams recording may delay to finish on low powered servers when two channels audio is recording
Symptoms
All the CPU cores are loaded to 100% while a number of streams are recording simultaneously, and the recordings are finished with a big delay when streams stop
5. H264 WebRTC stream recording may be corrupted when streaming from Adroid Firefox¶
Stream is playing normally via WebRTC, but recording may be corrupted when publishing H264 stream from Android Firefox on some devices
Symptoms
Recording file has a small size and cannot be played or picture seems corrupted when stream is published from Android Firefox
Solution
a) use VP8 to publish stream from Android Firefox
b) use Chrome or other browser to publish stream from this device
6. Some Android devices may forcefully publish WebRTC H264 stream with High profile¶
Some Android devices may publish WebRTC H264 stream with High profile even if there is no such profile in SDP while establishing WebRTC connection
Symptoms
High profile is displayed in MP4 recording file data
Solution
If there are any problems playing stream recording files with High profile, re-encode those files using ffmpeg for example, running it after recording is finished by on_record_hook.sh
script
7. Java machine may not initialize all the necessary modules in time after the server starts¶
The first record after the server start may be corrupted if Java machine does not initialize all the necessary modules in time
Symptoms
A long freeze in the first recording file after server start
Solution
a) update WCS to build 5.2.1105
b) if the build 5.2.1105 and newer is already used, make sure WebRTC stack modules pre-initilization on server startup is enabled
8. Webm recording files cannot be played in iOS Safari¶
Symptoms
File download starts instead of playback when clicking Webm recording file link
Solution
Download recording file to device and play it in local player application
9. Recording file is not created when publishing RTMP audio only stream with default settings¶
Symptoms
Audio only RTMP stream published to WCS is not recording
Solution
Use SDP settings file flash_handler_publish.sdp
without video track
v=0
o=- 1988962254 1988962254 IN IP4 0.0.0.0
c=IN IP4 0.0.0.0
t=0 0
a=sdplang:en
m=audio 0 RTP/AVP 97 8 0 102 103 104 105 106 107 108 109 110
a=rtpmap:97 SPEEX/16000
a=rtpmap:8 PCMA/8000
a=rtpmap:0 PCMU/8000
a=rtpmap:102 mpeg4-generic/48000/1
a=rtpmap:103 mpeg4-generic/44100/1
a=rtpmap:104 mpeg4-generic/32000/1
a=rtpmap:105 mpeg4-generic/24000/1
a=rtpmap:106 mpeg4-generic/22050/1
a=rtpmap:107 mpeg4-generic/16000/1
a=rtpmap:108 mpeg4-generic/12000/1
a=rtpmap:109 mpeg4-generic/11025/1
a=rtpmap:110 mpeg4-generic/8000/1
a=sendonly
10. Stream recording will start only when at least one key frame is received¶
Symptoms
Recording file is not created when starting recording (via REST API for example), there are no errors in server logs, the stream is already playing in browsers (until viewers reconnect)
Solution
Provide periodic key frames sending: for WebRTC streams with the following parameter
for RTMP streams with appropriate encoder settings
11. VLC player skips gaps when playing a stream pulled from multirecording in MKV container¶
Symptoms
Gaps in a pulled stream are skipped when playing it in VLC
12. Provide a regular keyframe interval when WebRTC stream is recording to singe MKV file and to MKV multirecorder simultaneously¶
Symptoms
VLC jumps to time of stream addition to multirecorder when playing a single stream MKV recording
Solution
Provide periodic key frames sending for WebRTC streams with the following parameter
13. G722 audio can be a shortly distorted when pulling a stream file from MKV multirecording¶
Symptoms
If the stream with G722 audio was removed and then added again to the multirecorder, a short audio distortions can be heard when the stream was added to multirecorder
Solution
Use Opus audio codec
14. PCMA audio may finish earlier than video after stream removing from multirecorder¶
Symptoms
If the stream with PCMA audio was removed from multirecorder, audio finishes earlier than video when playing the pulled stream file
Solution
Use Opus audio codec