English
English
English
  • SWARM Documentation
  • What's new?
    • Version 2024.2
    • Version 2024.1
    • Version 2023.3
      • Update 1
  • SWARM in a nutshell
    • SWARM Perception Platform Overview
  • Quick start guide
    • P101, P401 or OP101
      • P101 - Perception Box
      • P401 - Perception Box
      • OP101AC - Outdoor Perception Box
      • OP101DC - Outdoor Perception Box
    • Virtual Perception Box
      • System requirements
      • Install VPX Agent on NVIDIA Jetson (Jetpack 4.6)
      • Install VPX Agent on NVIDIA Jetson (Jetpack 5.1.2)
      • Install VPX Agent on X86/NVIDIA Server
  • Solution areas
    • Traffic Insights
      • Set-up Traffic Counting
      • Set-up Traffic Counting with speed estimates
      • Set-up Intersection Insights
    • Parking Insights
      • Set-up Barrierless Parking
      • Set-up Barrierless Parking with ANPR
        • Set-up guide and recommendations - ANPR
      • Set-up Single Space/Multi Space Parking
        • Standard examples
    • Advanced Traffic Insights
      • Set-up Adaptive Traffic Control
      • Set-up Journey Time & Traffic Flow
        • Set-up guide - Installation
        • Technical concept
      • Set-up Queue Length Detection
    • People Entry/Exit counting
  • SWARM Control Center
    • Devices
      • Camera & Device Monitoring
      • Camera Configuration
        • Scenario Configuration
          • Models
          • Calibration support
          • Camera settings
        • Rule Engine
          • Use Case Examples for Rule Engine
      • Device Health
    • Data Analytics
      • Creation and organization of dashboards
      • Dashboard overview & Widget creation
        • Traffic Scenario
        • Parking Scenario
        • Generic Scenario
    • Data Integration
      • Data Analytics API (REST API)
      • Raw event data with Custom MQTT server
      • SCC API
    • Administration
      • Monitoring Alerts
      • License Management
      • User Management
  • Test & Performance measurements
    • Benchmarks
      • How do we measure Performance?
    • White paper for use cases
      • Traffic Counting
      • Barrierless Parking and ANRP
  • Useful knowledge
    • 🚒Troubleshooting Guidelines
    • Network Requirements
    • Browser Compatibility SCC
    • Our Object Classes
    • Number Plate Area Code
  • Guidelines
    • How to access the debug output?
    • How to use Azure IotHub as Custom Broker
    • VPX
      • Upgrade IotEdge from 1.1 to 1.4
      • Upgrade Jetpack from 4.4.1 to 4.6.0
  • Getting Support
    • Get in touch
    • FAQs
Powered by GitBook
On this page
  • Event Triggers
  • Event Types
  • Event Triggers Details
  • Global Settings
  • Configuration of the Event Triggers

Was this helpful?

Export as PDF
  1. SWARM Control Center
  2. Devices
  3. Camera Configuration

Scenario Configuration

Configure your scenario according to your covered use cases

PreviousCamera ConfigurationNextModels

Last updated 1 year ago

Was this helpful?

Now, as you see your camera, you have the option to configure it. This is where the magic happens!

As SWARM software is mostly used in dedicated use cases, you can find all information for a perfect set-up in our Use Cases for , and

In the configuration, you can select the best for your use case as well as configure any combination of different event triggers and additional features to mirror your own use case.

Event Triggers

Each event trigger will generate a unique ID in the background. In order for you to keep track of all your configured types, you are able to give it a custom name on the left side panel of the configuration screen. --> This name is then used for choosing the right data in Data Analytics.

Please find the abbreviation and explanation of each event type below.

Event Types

We provide templates for the three different areas in order to have everything set for your use case.

  • Parking events --> Templates for any use case for Parking monitoring

  • Traffic events --> Templates for use cases around Traffic Monitoring and Traffic Safety.

  • People events --> Templates for using the People Full Body or People Head model.

This will support you to easier configure your scene with the corresponding available settings. You can find the description of the available Event Triggers and the individual available Trigger Settings below.

Event Triggers Details

Counting Lines will trigger a count as soon as the center of an object crosses the line. While configuring a CL you should consider the perspective of the camera and keep in mind that the center of the object will trigger the count.

The CL is logging as well the direction the object crossed the line in IN and OUT. You may toggle IN and OUT at any time to change the direction according to your needs. On top a custom name for IN and OUT direction can be configured. The custom name for direction can then be used as segmentation in Data Analytics and is part of the event output.

Per default, a CL only counts objects once. In case each crossing should be counted there is an option to enable events for repeated CL crossings. The only limitation taken there is that only counts will be taken into consideration if they are 5 seconds apart from each other.

Available Trigger Settings: ANPR, Speed Estimation, Events for repeated CL crossing

You can enable the Speed Estimates feature as a specific trigger setting with a Counting Line in the left side bar. This action will add one additional line that can be used to configure the distance between in your scenario. For best results, use a straight distance without bendings.

RoIs are counting objects in the specified region. This type also provides as well the Class and Dwell Time, which tells you how long the object has been in this region.

Depending on the scenario type we can differentiate between 3 types of RoIs. For those 3 types we are offering predefined templates described below:

Single Space Parking RoI
Multi Space Parking RoI
Generic RoI

Event Trigger

Time

Time

Time or Occupancy

Type

Parking

Parking

People & Traffic Events

Default number of objects

1

5

1

Color

Zones are used for OD - Origin - Destination. Counts will be generated, if an object moves through OD 1 and afterwards through OD 2. For OD at least two zones need to be configured.

The first zone the object passes will be the origin zone and the last one it moved through the destination zone.

A VD covers the need of having 3D counting lines. The object needs either to move into the field and then vanish or appear within the field and move out. Objects appearing and disappearing within the field, as well as objects passing the field are not counted.

The Virtual Door is designed for scenes to obtain detailed entry/exits count for doors or entrances of all kinds.

Virtual Door Logic - how it works

The logic for the Virtual Door is intended to be very simple. Each head or body is continuously tracked as it moves through the camera's view. Where the track starts and ends is used to define if an entry or exit event has occurred.

  • Entry: When the track start point starts within the Virtual Door and ends outside the Virtual Door, an in event is triggered

  • Exit: When the track start point starts outside the Virtual Door and ends within the Virtual Door, an out event is triggered

  • Walk by: When a track point starts outside the Virtual Door and ends outside the Virtual Door, no event is triggered

  • Stay outside: When a track point starts inside the Virtual Door and ends inside the Virtual Door, no event is triggered

Note: There is no need to configure the in and out directions of the door (like (legacy) Crossing Lines) as this is automatically set.

Global Settings

You can enable the ANPR feature with a Counting Line, which will add the license plate of vehicles as an additional parameter to the generated events. When enabling the ANPR feature, please consider your local data privacy laws and regulations, as number plates are sensitive information.

The Image Retention Time can be manually set. After this time, any number plate raw information as well as screen captures will be deleted.

In the Global Settings section, you have the option to add focus areas. A focus area will define the area of detection on the frame. So in case focus areas are defined, detections will only be taken into consideration in these corresponding areas. If a focus area is configured, the areas will be shown on the preview frame and in the table below. In the table you have the option to delete the focus area.

Attention: When a focus area is drawn, the live and track calibration will only show detections and tracks in these areas. So before focus areas are drawn check the track calibration in order to see where the tracks are on the frame to not miss essential detections in the focus area definition.

In the configuration, there are two trigger actions to choose from. Either a time or an occupancy change, depending on the use case.

In the Global Trigger settings you can adjust the RoI time interval.

The RoI time interval is used accordingly depending on the chosen trigger action:

  • Time --> The status of the region will be sent at the fixed configured time interval.

  • Occupancy --> You will receive an event if the occupancy state (vacant/occupied) changes. The RoI time interval is a pause time after an event was sent. This means that the occupancy change will not be checked for the configured time interval and you will receive max. one event per time frame. The state is always compared with the state sent in the last event.

At the raw track mode an event will be generated as soon as the object is leaving the camera frame. At this event the exact track of the object will be retrieved. The track will be gathered in X/Y coordinates of the camera frame.

Configuration of the Event Triggers

To create your own solution, select a model for your solution and then place your type (or select raw tracks mode).

When a type is active, left-click and hold the white circles to move the single corner points. You can create any tetragon (four-sided polygon). To move the entire type, left-click and hold anywhere on the type.

dark green

purple

light green

Learn more about the .

Please consider our Use Case specification to properly use this feature. The feature is especially available for .

You can enable the Journey time feature in the Global Settings on the left side bar. This feature generates journey time and traffic flow data. This setting is needed for . Find more technical details on data which will be generated in following section:

Raw Tracks should only be used in case you decide for the advanced set up with a

Virtual Door logic
Barrierless Parking Use Case
Advanced Traffic Insights
Technical concept
custom MQTT connection.
💚
💜
💚
Traffic Insights
Parking Insights
Advanced Traffic Insights
.
model
Scenario Configuration Overview
Event Name
Toggle for repeated CL crossings
Enable Speed Estimation
Counting line with speed estimation enabled
Configure Origin/Destination Zones
Raw Tracks