Scenario Configuration
Configure your scenario according to your covered use cases
Now, as you see your camera, you have the option to configure it. This is where the magic happens!
Scenario Configuration Overview
As SWARM software is mostly used in dedicated use cases, you can find all information for a perfect set-up in our Use Cases for Traffic Intelligence, Parking Monitoring and People Entry/Exit Counting. ​
In the configuration, you can select the best model for your use case as well as configure any combination of different event triggers to mirror your own use case.

Models

The model reflects the engine the Perception Box works with.
Model Selection
Find a short description for each model here. In order to decide which model to use for your use case, please refer to the dedicated use case in the "Use Case" section dedicated to Traffic or Parking.
Traffic & Parking
Parking (Single-/Multispace)
Parking Fisheye
People Full Body
People Head
This model detects vehicles, riders and people in highly dynamic scenes (e.g. urban traffic or highways where objects move fast).
This model detects vehicles, riders, and people in low dynamic scenes (for example parking, where vehicles do not move or move slowly). As this model is analyzing the video with a higher resolution it can detect objects that are further away from the camera. This requires more computation and is therefore currently only recommended for Single- and Multispace detection use cases.
This model detects vehicles, riders and people in low dynamic parking scenes when using a fish eye camera. It does not work with urban traffic or fast traffic scenes.
This model detects a full person/body and is ideal for detecting, tracking and counting people when they are further away (>5m) from the camera.
This model detects a person's head and is ideal for detecting, tracking and counting people when they are closer (<6m) to the camera.
​
While detecting and tracking people, we never do any face recognition. There will be no sensitive personal information being processed at all. Do not hesitate to contact us with any doubts or questions about this topic.
When changing the Model Type from a Parking or Traffic Scene to one of our people detection models, please consider, that this is only possible without having ANPR or a parking RoI configured. In order to be consistent, the system will inform you accordingly by displaying the following statement.
Invalid Model Configuration

Event Triggers

Each event trigger will generate a unique ID in the background. In order for you to keep track of all your configured types, you are able to give it a custom name on the left side panel of the configuration screen. --> This name is then used for choosing the right data in Data Analytics.
Event Name
Please find the abbreviation and explanation of each event type below.

Event Types

We provide templates for the three different areas in order to have everything set for your use case.
  • Parking events --> Templates for any use case for Parking monitoring
  • Traffic events --> Templates for use cases around Traffic Monitoring and Traffic Safety.
  • People events --> Templates for using the People Full Body or People Head model.
This will support you to easier configure your scene with the corresponding available settings. You can find the description of the available Event Triggers and the individual available Trigger Settings below.

Event Triggers Details

Counting Line
RoI (Region of Interest)
Origin Destination Zone
Virtual Door
Counting Lines will trigger a count as soon as the center of an object crosses the line. While configuring a CL you should consider the perspective of the camera and keep in mind that the center of the object will trigger the count.
The CL is logging as well the direction the object crossed the line in IN and OUT. You may toggle IN and OUT at any time to change the direction according to your needs. Per default, a CL only counts objects once. In case each crossing should be counted there is an option to enable events for repeated CL crossings. The only limitation taken there is that only counts will be taken into consideration if they are 5 seconds apart from each other.
​
Toggle CL Direction
Toggle for repeated CL crossings
Available Trigger Settings: ANPR, Speed Estimation, Events for repeated CL crossing
RoIs are counting objects in the specified region. This type also provides as well the Class and Dwell Time, which tells you how long the object has been in this region.
Depending on the scenario type we can differentiate between 3 types of RoIs. For those 3 types we are offering predefined templates described below:
Single Space Parking RoI
Multi Space Parking RoI
Generic RoI
Event Trigger
Time
Time
Time or Occupancy
Type
Parking
Parking
People & Traffic Events
Default number of objects
1
5
1
Color
dark green
πŸ’š
​
purple
πŸ’œ
​
light green
πŸ’š
​
Zones are used for OD - Origin - Destination. Counts will be generated, if an object moves through OD 1 and afterwards through OD 2. For OD at least two zones need to be configured.
The first zone the object passes will be the origin zone and the last one it moved through the destination zone.
​
Configure Origin/Destination Zones
A VD covers the need of having 3D counting lines. The object needs either to move into the field and then vanish or appear within the field and move out. Objects appearing and disappearing within the field, as well as objects passing the field are not counted.
Learn more about the Virtual Door logic.
The Virtual Door is designed for scenes to obtain detailed entry/exits count for doors or entrances of all kinds.

Virtual Door Logic - how it works

The logic for the Virtual Door is intended to be very simple. Each head or body is continuously tracked as it moves through the camera's view. Where the track starts and ends is used to define if an entry or exit event has occurred.
  • Entry: When the track start point starts within the Virtual Door and ends outside the Virtual Door, an in event is triggered
  • Exit: When the track start point starts outside the Virtual Door and ends within the Virtual Door, an out event is triggered
  • Walk by: When a track point starts outside the Virtual Door and ends outside the Virtual Door, no event is triggered
  • Stay outside: When a track point starts inside the Virtual Door and ends inside the Virtual Door, no event is triggered
Note: There is no need to configure the in and out directions of the door (like (legacy) Crossing Lines) as this is automatically set.
​

Trigger Settings

ANPR (automatic number-plate recognition)
Speed Estimates
RoI Trigger
Raw Tracks
You can enable the ANPR feature with a Counting Line, which will add the license plate of vehicles as an additional parameter to the generate events. When enabling the ANPR feature, please consider your local data privacy laws and regulations, as number plates are sensitive information.
The Image Retention Time can be manually set. After this time, any number plate raw information as well as screen captures will be deleted.
Please consider our Use Case specification to properly use this feature. The feature is especially available for Barrierless Parking Use Case.
​
ANPR Trigger Settings
You can enable the Speed Estimates feature as a specific trigger setting with a Counting Line. This action will add one additional line that can be used to configure the distance between in your scenario. For best results, use a straight distance without bendings.
​
Enable Speed Estimation
​
Speed Estimation Lines
In the configuration, there are two trigger actions to choose from. Either a time or an occupancy change, depending on the use case.
In the Global Trigger settings you can adjust the RoI time interval.
The RoI time interval is used accordingly depending on the chosen trigger action:
  • Time --> The status of the region will be sent at the fixed configured time interval.
  • Occupancy --> You will receive an event if the occupancy state (vacant/occupied) changes. The RoI time interval is a pause time after an event was sent. This means that the occupancy change will not be checked for the configured time interval and you will receive max. one event per time frame. The state is always compared with the state sent in the last event.
At the raw track mode an event will be generated as soon as the object is leaving the camera frame. At this event the exact track of the object will be retrieved. The track will be gathered in X/Y coordinates of the camera frame.
Raw Tracks should only be used in case you decide for the advanced set up with a custom MQTT connection.​
​
Raw Tracks
​
​
​
​
​
​

Configuration of the Event Triggers

To create your own solution, select a model for your solution and then place your type (or select raw tracks mode).
Configure Event Types
When a type is active, left-click and hold the white circles to move the single corner points. You can create any tetragon (four-sided polygon). To move the entire type, left-click and hold anywhere on the type.
​
Export as PDF
Copy link
On this page
Models
Event Triggers
Event Types
Event Triggers Details
Trigger Settings
Configuration of the Event Triggers