Create the LiveEvent Preview IP access control object. down to the ip range for your on-premises live encoder, laptop, or device that is sending re-use the same range here for the sample, but in production you can lock this LiveEventInputAccessControl liveEventInputAccess = new() this will control the IP that the encoder is running on and restrict access to only that encoder IP range. Create the LiveEvent input IP access control object Note: When creating a LiveEvent, you can specify allowed IP addresses in one of the following formats: While operating this tutorial, it is recommended to start out using OBS Studio before moving to another encoder. When broadcasting to a live event, please use one of the verified on-premises live streaming encoders. 6) Understand the advanced options such as low latency, and live transcription/captioning support. A live output can be considered as the "tape recorder" and the Asset is the tape that is inserted into it for recording. 5) Understand the differences between a LiveOutput and the Asset that it records to. 4) Understand the available Standby mode and how it differs from the Running Mode.
#DVR CLIENT V3.5.3 HOW TO#
3) Understand how to use long-running async operations 2) Understand the different live event types, pass-through and encoding 1) Understand the billing implications for the various states Understand the concepts of what a live event and a live output is in AMS first! Creating the LiveEvent - the primary object for live streaming in AMS. WriteLine( "TIP: Make sure that you have filled out the appsettings.json or. AddEnvironmentVariables() // parses the values from the optional. AddJsonFile( "appsettings.json ", optional: true, reloadOnChange: true) You can create this ".env" file by saving the "sample.env" file as ".env" file and fill it with the right values.ĬonfigWrapper config = new( new ConfigurationBuilder()
#DVR CLIENT V3.5.3 CODE#
Same code will work in VS Code, but VS Code uses also launch.json to get the. env file which should be in the root folder (same folder than the solution. If Visual Studio is used, let's read the. Public static async Task Main( string args) Private const bool UseInteractiveAuth = false Set this variable to true if you want to authenticate Interactively through the browser using your Azure user account This can also be created earlier after step 5 if desired. 11) Get the URLs for the HLS and DASH manifest to share with your audience 10) Create a new Streaming Locator on the recording Asset object from step 5. Set up OBS studio and start the broadcast. 9) Get the ingest RTMP endpoint URL for use in OBS Studio. 8) Get the preview endpoint to monitor in a player for DASH or HLS. 7) Start the Live Event - this can take a little bit. Asset (which is like the tape in the recorder). 6) Create a Live Output, which acts as the "recorder" to record into the 5) Create an Asset to be used for recording the live stream into 4) Create the Live Event without starting it or encoding channel type and size (720p or 1080p) 3) Configure the Live Event object with your settings. 2) Set up your IP restriction allow objects for ingest and preview 1) Create the client for AMS using AAD service principal or managed ID The workflow for the sample and for the recommended use of the Live API: Preset : Low-latency Quality or Performance (NVENC) or "veryfast" using x264 Keyframe Interval : 2s, or 1s for low latency Bitrate: 2500 Kbps (or something reasonable for your laptop) Encoder: NVIDIA NVENC (if avail) or x264 This sample assumes that you will use OBS Studio to broadcast RTMP Azure Media Services Live streaming sample