From 9ccdc43a7d07e4c6eeb4c98a485800aaa7963b15 Mon Sep 17 00:00:00 2001 From: Hadi Adineh <122263902+hadi-adineh-ascs@users.noreply.github.com> Date: Tue, 12 Sep 2023 18:26:53 +0200 Subject: [PATCH 1/8] Added more infos and pictures to README.md More information and about sub-libraries and maintainers has been added into the readme file. It also, somehow enhanced to be more readable and categrized. Signed-off-by: Hadi Adineh <122263902+hadi-adineh-ascs@users.noreply.github.com> Signed-off-by: Hadi Adineh --- profile/README.md | 242 +++++++++++++++++++++++++++++++++++++++------- 1 file changed, 209 insertions(+), 33 deletions(-) diff --git a/profile/README.md b/profile/README.md index 86e5115..68c9a0e 100644 --- a/profile/README.md +++ b/profile/README.md @@ -22,97 +22,273 @@ Learn more about the ENVITED research cluster of the Automotive Solution Center We are looking forward to welcome you as member of our community! -## Sub-Libraries +# OpenMSL Sub-Libraries + + [SL1 - Perception Sensor Models](#sl1---perception-sensor-models) + + [SL2 - Traffic Participant Models](#sl2---traffic-participant-models) + + [SL3 - Scenario Data](#sl3---scenario-data) + + [SL4 - Static Environment Data](#sl4---static-environment-data) + + [SL5 - Tooling](#sl5---tooling) + +
-### SL1 - Perception Sensor Models +--- +
+ +## SL1 - Perception Sensor Models This sub-library is a collection of [OSI](https://github.com/OpenSimulationInterface/open-simulation-interface) compliant sensor models according to the [OSMP](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) specification including a template repository demonstrating the [Credible Simulation Process](https://setlevel.de/assets/forschungsergebnisse/Credible-Simulation-Process-v1.0.pdf) by running full scale [SSP](https://ssp-standard.org/) based co-simulations in the CI pipeline. Initiated: 2022-07-25 -#### Maintainer +### Maintainer + + - [![avatar](https://images.weserv.nl/?url=avatars.githubusercontent.com/u/78017112?v=4&h=50&w=50&fit=cover&mask=circle&maxage=7d) Lukas Elster](https://github.com/LukasElster) (FZD TU Darmstadt) + + - [![avatar](https://images.weserv.nl/?url=avatars.githubusercontent.com/u/27010086?v=4&h=50&w=50&fit=cover&mask=circle&maxage=7d) Clemens Linnhoff](https://github.com/ClemensLinnhoff) (Persival GmbH) + + - [![avatar](https://images.weserv.nl/?url=avatars.githubusercontent.com/u/122266565?v=4&h=50&w=50&fit=cover&mask=circle&maxage=7d) Jürgen Wille](https://github.com/FM-juergenW) (FrontMod GmbH) + +### Repositories + + - #### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) + + Enter a short description of the model. + What is the purpose of the model? + What is the general modeling approach? + What inputs does the model need and what outputs does it generate? + + < Eye-catcher Image > + + + + more info: [click here](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) +
+ +- #### [sl-1-1-reflection-based-radar-object-model](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model) + + + + This model is a Reflection Based Radar Model based on the [Modular OSMP Framework](https://gitlab.com/tuda-fzd/perception-sensor-modeling/modular-osmp-framework) by FZD. + It is a highly parameterizable sensor system model including detection calculation and object tracking simulation. + The model received radar reflection data calculated in a simulation tool beforehand e.g. with ray tracing. + The model outputs are radar detections and detected moving objects. + + + + more info: [click here](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model) +
+ +- #### [sl-1-2-reflection-based-lidar-object-model](https://github.com/openMSL/sl-1-2-reflection-based-lidar-object-model) + + The current version of the model is build on the enhancements to the Open Simulation Interface from the publicly funded SETLevel project. + It is therefore dependent on the non-standard [SL OSI](https://gitlab.setlevel.de/open/osi) and not [ASAM OSI](https://github.com/OpenSimulationInterface/open-simulation-interface). + + + + This is the FZD Reflection Based Lidar Model based on the FZD OSI Sensor Model Packaging Framework. + It is a highly parameterizable sensor system model including detection calculation and object tracking simulation. + The model gets lidar reflection calculated in a simulation tool beforehand e.g. with ray tracing. + The model outputs are lidar detections and detected moving objects.
+ + + + more info: [click here](https://github.com/openMSL/sl-1-2-reflection-based-lidar-object-model) +
+ +- #### [sl-1-3-object-based-generic-perception-object-model](https://github.com/openMSL/sl-1-3-object-based-generic-perception-object-model) + + + + This model is a highly parameterizable generic perception sensor and tracking model. + It can be parameterized as a Lidar or a Radar. + The model is based on object lists and all modeling is performed on object level. + It includes typical sensor artifacts like soft FoV transitions, different detection ranges for different targets, + occlusion effects depending on the sensor technology as well as simulation of tracking behavior. + The model output are object lists for OSI SenorData moving objects. + + The architecture of the model as well as the parameterization structure are designed to be as generic as possible + to fit both radar and lidar sensors to utilize similarities in signal propagation and signal processing in both technologies. + This way, the model can be parameterized to model different kinds of lidar and radar sensors. + To give an example: You can set an irradiation pattern for the modeled sensor. + Depending on the sensor technology this can either be an antenna gain pattern for radar or a beam pattern for lidar. + + + + more info: [click here](https://github.com/openMSL/sl-1-3-object-based-generic-perception-object-model) +
+ +- #### [sl-1-4-object-based-camera-object-model](https://github.com/openMSL/sl-1-4-object-based-camera-object-model) + + This model is a parameterizable object based video perception sensor and tracking model using the interface OSI. + The model was developed in the project SetLevel by Bosch. The model should simulate some basic video typical effects in a phenomenological way. + The "object based camera object model" is based on object lists and all modeling is performed on object level. The model output are object lists for OSI SenorData moving and stationary objects. + + The outer layer of the model is the OSI Sensor Model Packaging (OSMP). + It specifies ways in which models using the Open Simulation Interface (OSI) are to be packaged for their use in simulation environments using FMI 2.0. + For more detailed information see the official documentation. + + + + more info: [click here](https://github.com/openMSL/sl-1-4-object-based-camera-object-model) +
+ +- #### [sl-1-5-sensor-model-testing](https://github.com/openMSL/sl-1-5-sensor-model-testing) + + This repository contains FMUs for automated testing, verification and validation for open-source perception sensor models. -- [Lukas Elster](https://github.com/LukasElster) (FZD TU Darmstadt) -- [Clemens Linnhoff](https://github.com/ClemensLinnhoff) (Persival GmbH) -- [Jürgen Wille](https://github.com/FM-juergenW) (FrontMod GmbH) + Further information can be found in [the repository](https://github.com/openMSL/sl-1-5-sensor-model-testing), the READMEs of the respective folders, and the documentation below: + - FMU collection for sensor model testing: + - [OSI Check FMU](src/osi-field-checker/)
+ This FMU checks if fields are missing in a received SensorData. + - [SRMD Validator](src/srmd-validator/)
+ This python script looks for SRMD files and validates them.
-#### Repositories +
-- [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) -- [sl-1-1-reflection-based-radar-object-model](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model) -- [sl-1-2-reflection-based-lidar-object-model](https://github.com/openMSL/sl-1-2-reflection-based-lidar-object-model) -- [sl-1-3-object-based-generic-perception-object-model](https://github.com/openMSL/sl-1-3-object-based-generic-perception-object-model) -- [sl-1-4-object-based-camera-object-model](https://github.com/openMSL/sl-1-4-object-based-camera-object-model) -- [sl-1-5-sensor-model-testing](https://github.com/openMSL/sl-1-5-sensor-model-testing) +--- +
-### SL2 - Traffic Participant Models +## SL2 - Traffic Participant Models This sub-library is a set of OSI compliant traffic participant models, which include pedestrian models, SSP based ALKS systems, automated road users and others to demonstrate closed loop simulations in combination with other sub-libraries utilizing open-source simulators such as [esmini](https://github.com/esmini/esmini). Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hello@envited.market) -#### Maintainer +### Maintainer - TBD - TBD - TBD -#### Repositories +### Repositories -- [sl-2-0-traffic-participant-model-repository-template](https://github.com/openMSL/sl-2-0-traffic-participant-model-repository-template) +- #### [sl-2-0-traffic-participant-model-repository-template](https://github.com/openMSL/sl-2-0-traffic-participant-model-repository-template) -### SL3 - Scenario Data +
+ +--- +
+ +## SL3 - Scenario Data This sub-library contains example scenario data following the [ASAM OpenSCENARIO](https://www.asam.net/standards/detail/openscenario/) standard to provide interpretations for legislative documents such as the UN Regulation No. 157 in order to discuss them in the community. In addition, the best practices to establish quality gates for scenario databases to clearly show the quality of scenario data are shown. Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hello@envited.market) -#### Maintainer +### Maintainer - TBD (BMW AG) - TBD - TBD -#### Repositories +### Repositories + +- #### [sl-3-1-osc-alks-scenarios](https://github.com/asam-oss/OSC-ALKS-scenarios) + + The here provided 15 concrete parametrized test scenarios are derived from the 6 subject areas analogous to Annex 5, Chapter 4.1-4.6 as an initial attempt to clarify the described set of functional scenarios. + + Each concrete scenario is complemented by a variation file to form a logical scenario, which then represents a set of concrete scenarios. By applying the parameter variation defined in the variation files to the parameters in the concrete scenarios, multiple concrete scenarios can be derived or generated to cover the required scenario space. + + The focus of the here provided scenarios is on securing the planning aspects of an "Automated Lane Keeping System". By extending the scenarios with environmental conditions (e.g. light, rain or wind) or references to e.g. 3D models, aspects of sensor and actuator technology could also be simulated and validated. + + more info: [click here](https://github.com/asam-oss/OSC-ALKS-scenarios) +
-- [sl-3-1-osc-alks-scenarios](https://github.com/asam-oss/OSC-ALKS-scenarios) +
-### SL4 - Static Environment Data +--- +
+ +## SL4 - Static Environment Data The German research project [GaiaX 4 PLC-AAD](https://www.gaia-x4plcaad.info/) develops quality metrics and tools to evaluate the successful integration of [ASAM OpenDRIVE](https://www.asam.net/standards/detail/opendrive) maps with e.g. [glTF](https://www.khronos.org/gltf/) 3D models and their respective material data extentions. Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hello@envited.market) -#### Maintainer +### Maintainer - TBD (BMW AG) - TBD - TBD -#### Repositories +### Repositories - In discussion -### SL5 - Tooling +
+ +--- +
+ +## SL5 - Tooling + This sub-library contains various tools to import, export, analyze and visualize co-simulation data. Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hello@envited.market) -#### Maintainer +### Maintainer - TBD (Persival GmbH) - TBD - TBD -#### Repositories +### Repositories + +- #### [sl-5-1-srmd-validator](https://github.com/openMSL/sl-5-1-srmd-validator) + + This python code is meant to be used in a CI pipeline, e.g. a GitHub Action. +It looks for SRMD files in the root directory of a repository, this repo is cloned into. +The found SRMD files are validated against the SRMD schema from [SSPTraceability](https://github.com/PMSFIT/SSPTraceability/). + + more info: [click here](https://github.com/openMSL/sl-5-1-srmd-validator) +
+ +- #### [sl-5-2-osi-field-checker](https://github.com/openMSL/sl-5-2-osi-field-checker) + + This FMU checks if fields are missing in a received SensorData. + It is meant to be used in a co-simulation connected to the output of the model under test. + It will output missing osi fields in the format of GitHub annotations, so that it can be used directly in a GitHub CI pipeline. + The image below shows an example of a failed pipeline due to missing OSI fields in the SensorData. + + OSI Field Checker in CI Pipeline + + more info: [click here](https://github.com/openMSL/sl-5-2-osi-field-checker) +
+ +- #### [sl-5-3-osmp-network-proxy](https://github.com/openMSL/sl-5-3-osmp-network-proxy) + + This Network Proxy FMU can receive SensorView and SensorData via TCP/IP using ZeroMQ. The received data is directly given to the FMU output. The proxy can also send SensorView oder SensorData received as FMU input via TCP/IP to a given IP address and port. + + more info: [click here](https://github.com/openMSL/sl-5-3-osmp-network-proxy) +
+ +- #### [sl-5-4-standalone-osi-trace-file-player](https://github.com/openMSL/sl-5-4-standalone-osi-trace-file-player) + + This mini application can read a binary ASAM OSI trace file (SensorData or SensorView) and send it step by step via TCP using ZeroMQ. + + more info: [click here](https://github.com/openMSL/sl-5-4-standalone-osi-trace-file-player) +
+ +- #### [sl-5-5-osi-trace-file-player](https://github.com/openMSL/sl-5-5-osi-trace-file-player) + + This [FMU](https://fmi-standard.org/) is able to play binary OSI trace files. + The folder containing the trace files has to be passed as FMI parameter _trace_path_. + The trace file player is build according to the [ASAM Open simulation Interface (OSI)](https://github.com/OpenSimulationInterface/open-simulation-interface) and the [OSI Sensor Model Packaging (OSMP)](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) examples. + + more info: [click here](https://github.com/openMSL/sl-5-5-osi-trace-file-player) +
+ +- #### [sl-5-6-osi-trace-file-writer](https://github.com/openMSL/sl-5-6-osi-trace-file-writer) + + This [FMU](https://fmi-standard.org/) is able to write binary OSI SensorData trace files. + The folder the trace files shall be written to has to be passed as FMI parameter _trace_path_. + The trace file writer is build according to + the [ASAM Open simulation Interface (OSI)](https://github.com/OpenSimulationInterface/open-simulation-interface) and + the [OSI Sensor Model Packaging (OSMP)](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) examples. -- [sl-5-1-srmd-validator](https://github.com/openMSL/sl-5-1-srmd-validator) -- [sl-5-2-osi-field-checker](https://github.com/openMSL/sl-5-2-osi-field-checker) -- [sl-5-3-osmp-network-proxy](https://github.com/openMSL/sl-5-3-osmp-network-proxy) -- [sl-5-4-standalone-osi-trace-file-player](https://github.com/openMSL/sl-5-4-standalone-osi-trace-file-player) -- [sl-5-5-osi-trace-file-player](https://github.com/openMSL/sl-5-5-osi-trace-file-player) -- [sl-5-6-osi-trace-file-writer](https://github.com/openMSL/sl-5-6-osi-trace-file-writer) + more info: [click here](https://github.com/openMSL/sl-5-6-osi-trace-file-writer) +
From bea55dd5b855b9af87d82b13a6c44de829a2bec5 Mon Sep 17 00:00:00 2001 From: Hadi Adineh Date: Mon, 18 Sep 2023 15:54:43 +0200 Subject: [PATCH 2/8] README updates: local avatars and sub libraries Signed-off-by: Hadi Adineh --- doc/img/avatar/ClemensLinnhoff.png | Bin 0 -> 5938 bytes doc/img/avatar/JuergenWille.png | Bin 0 -> 6069 bytes doc/img/avatar/LukasElster.png | Bin 0 -> 5600 bytes profile/README.md | 293 +++++++++++++++-------------- 4 files changed, 149 insertions(+), 144 deletions(-) create mode 100644 doc/img/avatar/ClemensLinnhoff.png create mode 100644 doc/img/avatar/JuergenWille.png create mode 100644 doc/img/avatar/LukasElster.png diff --git a/doc/img/avatar/ClemensLinnhoff.png b/doc/img/avatar/ClemensLinnhoff.png new file mode 100644 index 0000000000000000000000000000000000000000..7e486e63c397334401b612793aaef57a8f5651a9 GIT binary patch literal 5938 zcmV-27tQF2P)2!%oyiBy;)6!f-q-fin`I#xbhlO6lADkc17O?KRH zZEj-k>FT`Zj@zv{Qjy46DpO4VrQZMFx@r$V7bc*WvPX2YZt-AU?E&|=OSq1O5{d*;{4GP@Hic8nN+@;+sFIl{~kanSV(uj zLL`yAlFH?1yLthY1t}=ZjKkg#Pk36cM4+t>qFsy;xN#M>tzQZUorTz9uoRBEi?C_c zLO4*k=*-71XJa%sR6s5jzUFiJr}}#Omi_7gd_Mm}fk03p6bRpl#S%z_Jk%EKhnu+$ zOjpf^mHrag>&-*NW?jU%8Dp=9DZ(8M5MsL;J1kdWtLX}Om}tRc{R+4mX~JRk0@!QM zhYS5oNn#j;TrT)L0s4A+>v}moAO2bp_5uNaL?{wKEEGZ^lAtm@66VWhV)eXfu+Uxz zFAHtNxEf)%vp&K%>tT=UIt1J4V(W&b@G|`pzUIr}M-gD5g`K9$;k!W-ZflmoX_Y!Q zEuV#haUl?k`QUMR49qaW*!_|KzoRImOn~$Nfj|Vs;1F6G%i*+I4SMP`;b{0Jf~k}u z9CYDl_9eVY7d}>+@U>b^AFsxCixmjmqyrz*W$-dy3@<}R1s1tK9$^s??C-rEB9E9PST zl3B3UnhWzKGhw-OCN?bols-2TW=m&boyMmyS^gOu^cRsP=3=uBh3-6T(*6RbOFo9# zqG__FyB^X5%#rM6f(V-x@LjtYE-TewucZdZ)#|X*nvZo0rbB=B2QZocceqgYT(s3- zyL37>E}jnSMIXab{UaEu{S^jZd;l|z>DZ#T0IsB$C^r-E*j*5Cg-nT|heJSY&P_I} z`wu@Cf+?q_sSr%aNe?%Y z1HmrY5ApGX5Y!i4bq;04M5XQJ%Ko3wkLdytg2v6%a`NPld3Z3L#)hiu55Ci6N4RkP)>V z+YINxVdX4%>CeUW#2_e+WI=hn2v3@8@%U0b9$jj{NMj}BwS^FtWI|k#1Igih$g2x5 zP*Vt5RX!y2x!&RfU}eRkF4zrLn)6`0a-M3oI2lqW(_4(|7tTT`mX4PeX99%P&A>SR zGl&l%^)%CZ31XB6EZ1BegPV)mp%3(BAm16Vi&*5gU2=o8? zU)XE62EwXzI$$1#juc?{$U!LUilMBpz~G4r44tlq@ zRbsO0JV;8jAT7&=q>{eBl3uCAWVMIDr=O__afcI?);jf1ut{eDob}Za=4?r`q#J!S z@tHQvVY5{rtgbF5-mp5m)PD@**wk$BNgD!@P!)UzWOx|IDhp^{e1rwl{u}Dk{zO7I z2B-7@WhINWkOpCKD!rzX7V@DWL)cFs{q!oUo5%)1NK3OJFDIbNEQkq6Se*+Yfn3-h z3~yr%7^%}7S~>^LgihL-m4nk3x2U&S8!Q`QjZbC=lek41u$C`P6$C`zosC<~O{dzihCpl`q zO6&{>n<*_8OO=(bj?UKvGh-@9^dtn2?mqt=5C%*!IEb%rT!S6euD04o(4O-lQXEYn zIGjy)I|uTbJd7MK!5C>mSyx8oRtyD|74zE9MD-ymyF4nbTq?tCbRSH?os7MBkdpu| zl_ak;6U~Xi1Tr1E)QZ=WE7%j@LETb*6PUwUIVRwz|gf*U($f+Ql_km1?R3WFk zaG!jWE{rv1!;nhg(DqG`Q^)GBKS-LOR$G;cdwKi7$v*%-9aLDHL8X?hIyjeX{_UjQ zsM+a;SZf0WkjIK3*PLj(7FGT(xLc5fuA+4KZP3NC*&oAh#XKA?NQX=&Cy?H^AXCIX z3BsC$nEqDa{2xLHnOB-0w0l54pg^>bJJ!+!*8XA|j)Zzpn=M0cRR-$9JP~7M052M7 zQ8dwOcR1tDfhZLKm!|ZccSbuRXJ3ub}=q4Hv2rZXU}-veQUJ|T7{n9>qS#AL6@ zXVEOVcHs>CY}P_+)(1GB8VPQF0kU095oc+DPy@2&S_`mEV-BL|z>NX!;N&KOdng6B z_JyFp-Uz-LYA~NY4UU=%5Ms0f3FZX7-4;zTKIke=K$f2^0?qVLwss|25_Us6JOnO5 zOoA{v;jq3dEfvW9jx7vG&i6njU?xk90TB}jW3A-_15j|f(6HSFHmm00tHNZo<;T(; zwLq@DIYJCq!eYr>=*{{Fn+YbxdJXPp#F6NWac!SJ5@}QP*42cE`FgDS{BP8mv=C~j zi5kz1xDp?L2L&;x40S~WEnR0ROU(%}kkQJ(BM2_FU{!gId#<lh?e|#u&JyLIhtYnve{(E=R<$)OxS9wBYu+=N<3`wO=b+9 zTx=j=N8x1nRvZc1O5V!{6`pozjt;<$gk5-$9!=s;1nW=&&XMH_T|O6Sv?6qzK1!uE z1U^B$1)0B^Lm8l>SnG~}Aeuybk>;wByPNweUfk1Go(r|Q}J*qr5 zp>(SwYWzJ>O&L28;E9`wp%Bo_cy^%?JTj8sPyuojQuM zTj`FbghJL>fZoD5bm#2F*YRQaDsm^T$AzJTc6m--B2D>1JR`>}qV~$k-%k@hS2a&W zH08x*IgnN6({Rj%q?(y0>9}+t48>cF(L$5H`{oTu6bf2>-z_aRfe=gwlMx1{bC3af z@lQcWA5sZrg!Y5M;ZgM8zD}Je5$y-}Vc={94L$N&C(6iERgx(z!VohM$>|D;(^Y=z z(b>bQ!*cSH=rvCTH$>J%d4vXKHBEhHIMU?jmSmtMF$6gl`nZvwi2>?ZBB5Ay9!Bir zHiZG9taNcGBQ)hZ7@e@5zZYcE!cW35AX3Uo|F9AYnzm9JfZdg}{G2Ss@X0bfJX3=w z=a1nL<9h4MA+N{!%x==H$_VM?`TI+25@rHKLUTk$zgR)3R*sFJ^Kca| zB?Uu%itIUcrNM?0mD7EAqMD|64(_MMP}Wj0dbXOP2FjyF5RkEKj}O6vq)38Hr-7LT zK~W}sUm7~|5^z6nFX}?K;0$@Mv9TwrPAMe!Jr#s$zbq1Zw`z{i$DJyz48C9NZ?nSu zGG>lYc}ZnZkWPk3C&Oe}u2p1$U6~ASWd?XO#RUg5z|B2?_SjvxO(Ca}88}8xn7o#R z>^dX(uVaJI79Wfo1ay;v5&o_Hk+ePT#0}Ei(Evx((@jv0jL^*!tM(xtkM{!*70fsU z6TLs$e!rg)i3&ub0+G_*E=Rwj9}l0rKwD!4Ze&K{ZeBdv_U*#gal3F~mmkgtZUZ+v z8JEKYP`KF|SN8_v@}3h?9PzbB6b-kCbt~YvYCb%ceTD#?c}O+ zRXG@K2SRaf&vw-MJ0N3|0qQg331mdIVlyC2MwpKE{nGjlgfx=*LyPiGMwrx!L}C?) zi1ueGl@_yp4v#&>gSKlZ+GdW1AQ!r;{x}`zfkf&!5r*oh-Aew+&zk&{7HsGL9qVWP z8NQloIOt@8I$t~NHCYCKS_uw&S>skp7!J8^K$({nQf+n7Osl}~=p)sN&75Jf^}esP z2xQ9-P2+oE$f_{Ud!=$!Rt9PH74v$L@9TivE$eYHb_dQ!`XGQzV~F8GB-t3D*ux6B zHU`*ls*QBJ4OC_uQ0!%cKtnA!E}o6ZHA_*seIr_vf{<>bk1WSENU&ascJgyd<*4cr z6my2305WaT0khw$ktrZdW|;C~VB}=$N6E4rP7X(`*>Zw0L2F_d3f#<*?_z}%dt10` zE9pc=+%kt~S=dUvB}DZ4GcX$_Hm7wh>S$gk|}- zAD@7>#3bCvNJo2GI&Ou>peHQ@oRTbbmL}s0ec##L+i@b&4e=XS!$)T>4reA~bZl%~ zm%{M}A5$};nj;g2al!#dOa=LbKV{Wrg1HOt0!?aVJS+WBkfG}zR3OIH4qf+ z5)`q!FfcTvYWJdv88X!{z60`#%IdpScLL_9KV zvxGXJKjd5la$W7H*t!CglKadhyLA+F$6rb6CWjJw(Ua49Yn*JzrH4W z0QrDoc=8VPsW!853*E;Z|2L1zS9!1g!C`PkPtYxShA{ww_I9+!h9KHZA9l+=N3hXS z>^53Po1+mb+^kUGumLGn+6Xt&fZOspFrPOA9=htJE9zXky;W1af`;nN)_XW{<}Q-c zYS8h3J<%~~Fw#!KWB&NVWTcBV+}+jnKko&ZfU%i`_+Io-aK+#$M$rHC74Gw%qV>U3 z^eLWW>?NoH4B&Ro0h|f;L8PS~Y`&ZYOSO+-zhnlSm(PU5q8YGQI34=5>zgi^PHk2f zr^yvu-WNz9zGw_|L*(YoxP6a@yB!Mn2V`M?ay8C3-Ng0l?YMR8E*{(;f1cId)#30D zAto<6cz!Qald8us?C}y4^+Pi99D~o^fIsvM+<|Af**${3{>K=8`WgdIfm;p7(Mq1{ zSg1E5O!Wvr4c2sMGKr}XxUe^X2B|lK^}a;Z&Jc8R zhrr?vBPFLE?%NVkRnv--=kDQj%L6oB<={#iH|Hlos6cxAe#`0UImaddp>zN;7-m@*(3d5J^y-QY+bL;36#lrKP;6QFtT9^BZ!2h}^g5oczEZ7b%%jk4oS zuc7*jk?pV+SN8f7L^xVgqR~X2EY;2kYi7+rO;sZ}yoX@-jo{enc5L>DM_F|nuH2NO z?cN}6+#a~ldVApK{XLPx>H7m$^6+M_;#-Kv-atC?0z%4&{PDNo4?f2=`o%@{EcD4= zVc-#EjZ!Zc#F-l)PwB{m+V9MOBaFsgN1=<9gq63lb zWC1I+Pq0zP0PXi==Dk>aBTiuXi zV~S`aE$ksDe0--Tl~Wjf%^x{#HgH{`0nKTDMNC9IxQf^4>KO&AcLeuY14v9egUaJ9 zeD%%1g|@qcfBcsqCa-SsiUQRbJn+WcYwQ?DM`lRvlIlxOzJUc43$zsA|G6kzcm zLqUUt$;zWQAW`zLC$0q9Ij3+mGY3@xekc#zfiiy|l>2%h-@_5RtOpfTcUuv z(}ARNIM@YYyH7N(Tw}ix2%lE4dPY9@RUoEbJn5?iJ(3p{=dMd$wRH?pCID(;sBUP1 zrL`|S++&cP*MR!t7g2HWFb)NVB5kuhlI`sfYi$W%13iR$`r~LrD@saFASxsUjyArp zS8J_1rLVtH<5vfG_qx)qn9=%8f7GLwV4c5u2WySpp{s2U`;FU>pIwPFr03B}5o$gCGZ;RP>Tw*)Ff%q%!jjxJ3#Wv4UiPu$#6QF>?CuOv*O9k-6TFqpL-S} zDT*d_b}ZlIoq1-t015m)@AEw0=li_NKlp=h=$f;w?_992mFQO2XY^|;*R=Mfj~6V4 zPt->1ujWm*Uo9+|KUpxEK2{r-UYs(SXC}=n#Z&f;9pBXN|D7(4d3D#Ed8K>7w*JxL z@;Zz5WlYP<7;R=uRs(wTB6^DjgM`UyX31tjXEvdg_tQ%jR7+Mz{CLW2?V7c$JoEn@ zVA`_msy44IEtr=-)tb#1tX6bJ1M2x%)U%U}^mlQt?=(~A2T@I&WAy9*<7fMr8XHD4 zuR^C^z+|_g*#@#QV_s&~vi#YsW$oEn%c{@s9bnRAdQ`Qvv@~P3ey+CInU}k&H)t^| z&M`GULjUPj`cE}7+}_MgUnjE?V`ti#8t!7euZ{kbwe+{wGu(TMe);qpbu&fHK3HuVVFk#^MmCz$GMX5@4OCmYJCEX|>au|=S{0^|ZuBFam`1zM4z@7cQ;)i@3G2)Nnz1v~m*-MhR>bu5BvzYE zfCR*1W6rk5v~7cF`v!CN^*IrdZ~aaXlLqUzC-uu$B~a-t%m~Q5W|sNs^NjVKWT3eU z(?AEWSv36Dt6Ti-YuEVc%@=t4qMc>c2n$j|=JVaGO!Wy=8)GL*>8r~_cTWDz6+0`7 z)AY18(0#IrY1Jspt9JAPq?HWKO1VwjRvjQL*w=4pRyP0iKL}y@j)1&1VOV2QzrvJ& zsP#IA2D|BPtEIQD2z5t2H;rmO*s}BY0`ZS;zskq2UgdkwZE#tymMnB)9qnODJ%mm& z(N~j2OMV=~EfuV3$9d=a1#U_PC;K}X9XO52Fpq83#-b?Syu_>k%?pSl1msPLKl!a7 zCKi{zB_Wxx;K;_fZb=&3NUt2Mwj!JQ+&CtV7jsiH#ycw--o0SrM>j6=w{O10k8i%n z`&TdU+WHb#G!tB$8f0s{4^_)iTC-z0nU_F+bsol{Q(V@L@#=uY-YTdO2H$1oUX zF|An9uC1Y76#(fOjr~sC5UKy8-#i3SyW3+XLRrZX>&{)!CdPFArr z*ea#h!KKA9^rNTAi4W#jWj=E{HO9>i<^@E%vWafx!rd|V8$nDPZCMgHV#*)`%Vfp=Ow{I}Jz36TXEh65HS+TWYP5sK;uMsjo>WzrNEvIqQM`Js}0Wu-44v?~=WrX;-aPnw2>k~a(cR)xJzce?%4b31g z8^?LgHqSc(@a}~rzH3u+d8(I{ZYh_RLKYfwQB|cfU7I0b`7E}TFyB(Zctai|wb@Lz zmC_@yu|Y1xMF-J0d`4!BD40cHZ0q-C<(K-kt;fF(#C#iKI~*A?*iajmC@Rh;FFA_E z;Zr<2-OGy-*C%_pG1J9$bq`l(dwFhdkQa65xxRRw3lp6f9C!Le5esto+3M8on2}IB z-q)owS(8d%NjzOevGkND(O;KCR*WzHUI#edTq_FZm@U%q*0pUI?edppreC{gd=P}E z#OHT|IGTMzza(ikkeiuCZ9zKbbDgY-lx_{T^4z&LE)CUlbzELYPqIANh;^U|b8kJ` zQ)SH5XQ8f2VWKpS@se0Zi;1K?4L7mee_ZOnl$nNX<{C1Ysz_p@ByJmGC@+HkTm@YjL7YhNqAtpn@^CjQ zqJ5}I2&6ndh>EOOB0`Rklb4ObZfC)|x(%>k{Q}0JUbizNAS?Gm?2`EG*C8Y;CM#2O z3&biys7zC$>8!_grh)a|1_5Z~%Gn0C25Z?EtmpECl-)!Jn`c|Fc2%Nn%4J&mVp7WH zTyZRCvcm)-gx16()W^6{6X`-}m^;M^56YBBD2@}D3?=bVArzMtOGc#8mmLScunph< z`VwRt<}=;O#hrJ9^xP?@Te7gM_RVV)Bacz&jji)Y&~ww5w44Ln_u#<_eYXA5IElNCmHRxqte-U8xGMTjef zq0Zz>c8U^?P?w`5KR%NBnyPIOhict5N46pE4uR>s8|0&VK~xfBdNYIOC3N$8vi*Z8 zhzaKWi4x51m8^GHW0$hpIM>3a^vq>x{A(h~FU|t72Z@x$8?JHFE3 zU&)A>QHvSltcHSM1tsy}%=I0|(pkfXqqL^ld3|x1n+~;7^|C(L#;Q16yUZD5lgt7K zKw%Q&Maig2Q&Gtw$I2ZLF|;HGP#5h@Rj3z5l9g-$sm+b#OhYk>BW|4RJcV&YF=vXbw0q(#ssL$xk0fTIyU9E@WM9g$$^mliK_Jso zTJx&&jFy$LE*53DDu=0#3I+?}X;b=<( z$5DFhq9VwZ&Xh>$1urr@50L1-k7(C@ggfs+;j)JW_X9+^?5;L#cHaz_C5aR zI6w0kVTX4UCbDCHNkwj~yVtxG%_}g*l@I z>%iihxf|Y z1L-MBC4B$mcs}`O9DM9M?D+FP=IMvN&4H&L#&hqJ1bgixInIx$KxcB}Bd94X6v@}2 zF-u0wtJ@%R7AZCJ${m>7C3XwspB*6o{-BJ^i2N?A=QtJS!zsYy-#Ymr;6v!p5oe=7{`e|u`BJEPl$rmi0oDsGhbaM)+A@! z%4pikC;^NhOVxd zNjc2Iv0`+!r5GD4c|~0Dvdo<7C?67CpCQG|N!02P>7rO!qFfn4&J=}v$SiT8OZJg0 z0V_{W!fG<1GugI1)^7mv&kt<&Z4i-Py+Ma|bda@kU8rg@WQWMa(km)9P|Zen6&J1ZkkIWhM%#4pPP+u-{L+tqR494qa zL8?hcdn_G|EJd?bGQmp{Ih&`XBi^6>j7R~>l~SwV;?N1^8w)v|9YdDqK9OLt=&}VK zlMK~JCTgQSDVCX2oe)665hpUDL$Mn4asy>=Us}84dhd<9)l1^)*A7I7`yDqJSkP%O z&rUGg-NayBHtqQePA2-$ljuiZrl?tZ2rVL?)d3>4676EgEBy|VzVA_5;{(OA2U8U2 zMuvwpwa0!+gFU_g5!YKA?nzmQI~DOEWJCqg+18B3wv5JPM{C>sN){a9P%ibYEZxqG zgzDEU%2$-jXyv?F)?L|0EIKW#3o6dF9+f31jH5wL)XG4t4mm_wzyWgIcaZP7iU^Y zsm@Q3=JFJ^a+llFBI(Rjaw1VdslO|kl7VEA?ub2461L}QB4tp<9oj{; zFFWUDII0as&ULiYm6u9uia!mB9@3=l<^e^r8I{`^1{;CrdJwC(@fBZnvYk3CoJ(?#mFgT)<$vgvlnR(Q222FH5Z6 zdRXUpjJk>10V$!et$%oD9}$r52ZCtrObE!?C97D5CC+xX(;+KwN2cuAsd9Kn8AZz- z^_2xlqQLVog>DBaaX&<%vy`0cellESa=SPY?{bJ}XL*&sqjVLYCC#7YbBIVEH(Fbo zr6<;zHj5#&UzQ$`Ro1+@{iJ^0x-1#^;?a=9L^>XHFF1^&eOt9U?gn{AAU?Yv$h6M3 z4KgZaHZLIKecd!l`JBoKpe5Ozj@)>TM*33{;6anvn8t`;I>dW*#YI!)?M|j-AYtz= zLU%lg-_wubzxzo-4(7}q{y9t(N_t^Qc*pg?6@i{<2WPR(>A)Nt*Dn@WMSp1WWxa?Gs7x&-XU%u)0_Mw zKBW72$j)+@5a-=^?s<%Gnebs=hw%6D#m8TPcW5F32_^W%m639^ixJ%=$;>rbbzelc z`hs|_ivo2~vLa3H$db|gxz1vDebBQl)341Dqe*(lgj#QB*68?%%qnJbo?>b0qL@IA zC52KaFsHJW43wrbRFp<{nxokRsqk?o+xal*u42>OorrYXOMrN-AlX3zrKEy9-0=1e z!Y446BZ^F%1G5N7tl?pTo#E;%Mk+EGE=i>) zQ}%LY5aj{hnwe8cn3iO<8yc)0UjRsx_RT`8bp911JA%r+=GjV%R5bNm!9@f zJ1v7oh8M3hpt*`}>s>Ct@e@|AzAgQ8=@a$R>LU*X`Jz`WnT||2?8&6w#JEY%tN_hw z9G@gzX7!C9aQ$!o8?XQM-_TswNM^VfRjEOAm8Z}v5Pjw843=fcjuK0=OzUbHc13}n zq{?87_SsLc%#{Gigs-a)hxU5l?j4O^XcnG9dH6-kme*FNOg?|#IESKiP2T8P`%wB9r~BQs-GVtP?zSHH>1bMJEPy`S^uNB@&ofACLi zJoh$bWlh`OC@aiEwnW*<)1s+KO{5|rj8~a55+q$U&^W!&(I>g!b%B@t45jIM)`>eDo;*R-EF3-%S_w3O(w11l^c6w<~M?H zTvIyBAF6c5=T$lrE0=Ea>i7SiH-Gd?zWbwp<@FDL$yNIzuy z9wIBiT@LI|a$q1CeqLnzxRMdzN3y>U$$p-s1o@!wbZ76AdwBW@7hK)r@D3@!OHm?a zRfwm&2E`vErQ{3^-HX(8YA9%)qoTvWz`T^${0rALo8SBGK2KCFuKmf@^>^NU<^7*= z?cJYq^}T=M-M{-kyz}w@=7rZjrlENVFRAaEj7+NIqezwgJkd?2w)MqbHAN}hAEQD zM>syT%6Y@9Z_KT|_uGB+cKf>V-p{^$_3fWM|Jo1#mG^%3Iq&@4XT1DZpJ2Q27G)JD z@o^2NATplD5m0z^mP)7hsmrRXHfUjb=}qn-}w(g+`hi|(_ek-?Vo-+ z_m)7s@sm$Q3jYDM@j4}Cr|@zKCC*1dQ))V8V*WBjRujElWze}3DZLOO3gvs?Dcp8E zhSS53vhOc;;JnidSJy~9f<>jm^YBnOA|JolN@8={B@2^OcNpmyyMi)*K*~xtGiiVG z4}T|!yRYy6oa`5V@V`DWUj7kh)mJF5=|>rzPndfMO&K{f#72@YUMt1hg&0p)qCK4n zlbLa3*CTlDd;+H@c5-mHC(h21xOgWCNS2gSF0LVY_(YWxkgw8Q0)kk=q=6rO`dZ%Lz$tB3VkRq-BPt!R61-%s#&~ ztbh40e=iVsU*~l%xeTZ-E=_Bn|8!(P&GEuY@m>j3DE!Ek#!mC|CBaKzTn`epXQvp& zJsj~0;IAY&IEkRJR6-(i2#+fzB(9Rk)H*VY zTR*I980xHP8~@Mof46G#3ZE`gMiyl%6jRB;VK1bGhJTbA8u5$d+8ccB=%m~=6C00000NkvXXu0mjf&ZN;( literal 0 HcmV?d00001 diff --git a/doc/img/avatar/LukasElster.png b/doc/img/avatar/LukasElster.png new file mode 100644 index 0000000000000000000000000000000000000000..7f2806030582f47228d8dd3702f287e0abbe7f84 GIT binary patch literal 5600 zcmV<66(8z}P)53goL5YExd(Jq97|kZT$$jp5p4T%o z&v1Uf&+q$IaLABf>5>RIBP1f8mrN`u;MP@ESC*Y@D?fJd?%6~653ip+{*YID{jNmN z&?Xb}Zb(JE0-2EKNk23E*INJIx+MI@kz#)1b`iI}gT}J0UMA8r7FdK#vA;jZ=Sn00C#HL_h#S&SM3S0~wj9z7G4hhr@B| zJ8+#n3bvEqq|J_n_mVO6`YpK4e;>9}-i6J?ci=K_ESzV52T8e)xRBB_C49~~F|T3hZvr6|a9qgl$(Y3i97y>baBo~jaqezJIxd9Y;`gz3 z$}p_@=pFbkn*jH@@561*`v_j~5!O!~g(c(OfaBaTSU+1mHGvI2@g!+HKh_$`~t zx+f`45LV-WTnE1{)WB0Ox7L;WA?c>?Xa2sqemysUwGC=6kOa z7{w5|Tga;E2wgW9o=YYoJIVw6sw=Ge5_WAa1Hzb)>`%WG;w5^>50p@lsf7a6T`I@v zg5B6?JB2p=8YT?;Ggi{+uAB5aylL}+bKiq6ZQ7qQm)ETKu>RvwSn&Q^SUC1g*iIad zPba+ryN`xp(O5>s*Wfs91e|A$f%Cj^m^W@1!tLkb($PFxQ^2ahfIOqjfBX`Vq93@E zh|4OX6^n84)FFh}&u2F={N>+a`3JAUZv5*Aq>L4_WIVzaj6?GBkCC=!4z?|uhCL38 zv5(Simdi54+s;J*#lzAuuVDF@|H9hILn%SL1y@qgivr(i?l{V7qp>f>4}#h%T9?lz z6)Oy*3?+aAKNrH$3L$->go^UKlH5usksu=`2zFE6#)LQifU&Rq9!uzC!^wSeo#&$1 zXC==0uEs@wJ5+`_gA?xye&R-O5`9q<>wycw>yc-_5K%Vc;X7*-Ts|4WW;V}7e)EHcPA7wu_Yz5ZbmLfFd_?#+fPIA8i%qe_DN7n#XJ^5CE!8L=RwZl zz;6wy@!p>>WBALMJL(mLE}x9UzAI1?uo5>y?7@w3gDAlZvZRer?Fxb>BLpfk>COP~ zwt1q0zEk8p53vOAHFX#^&KreIwjWVK7z+m`Im{nJHWvG~21CMWBvl0LULu_K=6|MQ5xccLci5G5w{UZ9(DvV z7Vb+XVC{@Cl!`yW{^SU9r3My+-3x(Ol@T$8n8!W-A~ilHfX7w{)+GEVAUtR_a-2Ps z3!4dV;nR@Coy zhm$(t(Cm$eE`J9!C-*~p?L1DTM8nr+2K?48g4?p`SUYtz_EJp|b86Y3w}LQcaWTet zkA+;$r01y7cz!U1Ld<^(q85@#gy7X&#isQO;6wqr`r{E)kw;^f!#r@dZh&&TFEpt^ zXi5u3^PVkeW>m?ILu=kvH0Q)Xml+8yK`^FFkAx0ura)Le&BB8HrSU;_UMh(#41qe{uME;W+u(NiS`fvBk|)RMevdWc>vM|RQ{ zcrTp@R|--lb>@56K#=GSCdeyGU`H@hO-$p%pyR>F7n9C$39iv3CCiUMXYeAcWW78r^bKA%zL5y6ZY1j0>v zRt?sq0t*P4T0$llqoVvQLhTpBX~swj)X}&S=?FoBCsezGq0b0|At#!3EhD)UqX+}j zloLyF69vP*I2a0&pwCN$z9@r04x**xAWlVjW7Ep%@L#hK8`mvFsKY{B{_GGm3OVae za#~;Z6o%@HK)^t-Mj#2_gYda6F9K1~LsU{BR0N`=yS%XmTU?f7)udqvq+Y6?VuME+ ziksk#nr*(QqCggPWUdvA*wZH z5iY3S<_AeZGFmSbqxJkz6oh%g&1O36=Tkc09S^O93334x6q|!#Ebhf;Rk6a5D#IBN zZ&toRYCI3XYC5GH7n8%`FnuIaoaUinhd-`H*`wHdIU-jrfaCmG$aJ(pIYCLu-C7Hx zP#NWfVy_iA;I<6mtL9V7KOKK_Ux6ExAQVS3V7z=BN4EvT-DV~=m7ky@VH0$f<&aa# zS3H4uR+YgZq>2XvQt)#?)G`s&2^|WzYoDcUyh@%VE*(^tar#(tT*FsrHDZ1<+Y6^EC+;IhV zIxmB0ix2eI&p{!VLdhhZ=em~#l0llg6A8A{Q53Wu z;&gv-Q@y}V@@8Y@O6Ufh^jU-Rl!@v>9MP5?kNVUo9QIg^v;K~_9PWjx7+(l8wt$xw z1x3Mjn97cz@jx1qH?Bk3X6g(=J)y6>0EI%vD)K}35<$rQ)1BVjYaFpz39Gp6m_{t$PaTvJym-}{3bR=7yu^03VthcNh?sT>Ju;Vj;v97?lL_lLyfd<%QxfM<_BwpiAG3=G~!CCIvuFLCmAT zu8VSkdTRi>j~Al5q6|&P_Ck~z32{92PzeE0r-s5rwAh2RDSx;!fs zO$1}2I?Vdm%`t)1}R~5R55@s`n0GsFOXQNc5rCn_#Ab*l{oe{j^zC;ZCYG1eY0%Hmdd} zsvYJM2BZk4^Z95xR|xa59JHK20S&pYS|)u4hB6qnD~vrCgaGca60&MoKo}J)?qmVc zDlDc`N};pfd9|o3N`oXl2!=f&Xv^FLRjLp8ajv+&X)VqN+ksDU!%GRkr7#x?VkcY; zc0gma3-r7F(Mk#$52vHOG!KSzIglPqh3P~#nj|zbFbBmP8m+C8S&RWuO2t2b85A$K z7!cEsA;eE1EFelPs7zJl#rEE=DHZ_ ziWApL7$>}s{|aCJ)>Q^TwqJPaO8A36g-h+ufngCAa)66Y3K?A+IQfhMuXVHMB|wL`D$gVsaU4 zgF&8kGl^JO$$%8RNDV856~qdmr&^-dY9MZ?LdC94EQma9BPwVvNm)p3eHzUqM|Y#E z{1gn7dYjH4hKO4E0=E@NSU3xnNj~I~fspLk1m)2rh>MaTtGxoP$_ipJrfD1uGFX|x zAX14acL=35@1FzlLl8Nd)Ry#E{kJp-c6ih2hM}n-2EybJR8!x?%Z-Qhh9=&x3hZ6$0@XLJ-4$ z3d9OA7=(dnqI)x`QdfQw7vd>X(Z+PtS7~#jA&w1!C^!g>5n-s`6a@Z8Uuct3(Q>K? z@Ak4i*uZ4)FxZ_)? zg-`W{?f?x;xiK`##?gF2U2kqW)ESA8rP0iOCd*f>X0t zmlBL>s`i2m`m8XCV3KK=%7(r)7q#@AG8&C<(0tM?sD)mqW!2EDl`P0VgBb)8KN#d) zvPZuJgjJ+T2maME9H;t_<+d0{Hm{)u(GNao@z?n7GKoH2T*|fQGM}4{mxMEbE)H{gm5`60>$)sWA0X* z3w3~|`ZDg^{gQ>yvvKl#kY`~CmDUNU1fKyxCP#Wih{B_fD zENC5Wc6Z}{9s>gdH__H=L~BzssiUFw)ISAt#?8IITGrK zGxVJ!)WW3CAA=hOyYThHNBH*JNBH84uh67#K~K*ubQrs7y;gQj9aHfY3bxu?K;(DH zwO)H62n(Z-I(%PU6|CPHXv1wJyRq4TPLl~{qX{Mhot$2W?v4<rNizji_e=98i<8 z85iQbP!{ch5}HX)ZnnoMii46USDdEzC8Wr?FlQWcnTvyt^PsM}if_oTzM+4<`sxAN zTH4XwHGrF+-$zGBH#*vTVKSLvXlh|AznbmEp+*O6CJJXl%b5Y=}vSqr$nE%x0=XWSeUk!7F<4h z6a}uUP#Nn_ebiP+bGAa99S7yUZBXP6?>4ASTwYCclO$KU3bgZwMT3T5( zIy<^pIAb%De$S$z^a$ECN)&DK!D*W1xjUmE$%?0mF9Di;@zjUKLzNQ^d1e@dDM6?R zbHK%DZ|HfAh|~!g_U^yI-QfKESw=*+_9gF=K8g5IQ-Hgb%^>a;W)@Q1W=_AUCT!eh3#L zyueKfgCd=(InC<2+-*?riK6!*;BBWuxFZx=P6KL6&p=pp10 z?(W;TdGkwr{`pr09UZrS_isTAzS`T2PF)@5yG+27Qt0mNW9JfaT_wedkzB14%?cgb z+d2v6Fe;*aAtD!3W^RQxD;AnPF;HcOgP-UN)%jCstgHg(>N$4)*QvGW@A;fu>jC=u zDX=?kQwqL+r@#Nc!>ab_o%mffHGG*npIj#lLK^u0~Dxca=lJy zghHI!umZI)K9FX_K${%{T{gwY&JbwM9j97hKx0)Uv{YPM=ygYXKV_v`q`-Z2bqzf3 z?7Ugr+k5Zz-vnat)!%!2LSN6V;-0R7?>Y&hr|S-SDMj}7FhSXi4st*K=_9xh9e`>w z$?j+ovTmL|-R7JCRCAH}in3Ou)q>A0&cMk@AE7hcKHnq~p z>w9|tp3u|x#s820yG5INr?2mxNB`${^7?uPs>uCXZr%RJo#tlK1D&q*!N9=XI|Kc< uTDr}Bm2E8@xm{hiJ$eTI{`Rl6{{I4O8)J;%g8AzJ0000 ---- +---
## SL1 - Perception Sensor Models -This sub-library is a collection of [OSI](https://github.com/OpenSimulationInterface/open-simulation-interface) compliant sensor models according to the [OSMP](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) specification including a template repository +> This sub-library is a collection of [OSI](https://github.com/OpenSimulationInterface/open-simulation-interface) compliant sensor models according to the [OSMP](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) specification including a template repository demonstrating the [Credible Simulation Process](https://setlevel.de/assets/forschungsergebnisse/Credible-Simulation-Process-v1.0.pdf) by running full scale [SSP](https://ssp-standard.org/) based co-simulations in the CI pipeline. Initiated: 2022-07-25 - + ### Maintainer - - [![avatar](https://images.weserv.nl/?url=avatars.githubusercontent.com/u/78017112?v=4&h=50&w=50&fit=cover&mask=circle&maxage=7d) Lukas Elster](https://github.com/LukasElster) (FZD TU Darmstadt) - - - [![avatar](https://images.weserv.nl/?url=avatars.githubusercontent.com/u/27010086?v=4&h=50&w=50&fit=cover&mask=circle&maxage=7d) Clemens Linnhoff](https://github.com/ClemensLinnhoff) (Persival GmbH) - - - [![avatar](https://images.weserv.nl/?url=avatars.githubusercontent.com/u/122266565?v=4&h=50&w=50&fit=cover&mask=circle&maxage=7d) Jürgen Wille](https://github.com/FM-juergenW) (FrontMod GmbH) - + [![avatar](../doc/img/avatar/LukasElster.png) Lukas Elster](https://github.com/LukasElster) (FZD TU Darmstadt) + + [![avatar](../doc/img/avatar/ClemensLinnhoff.png) Clemens Linnhoff](https://github.com/ClemensLinnhoff) (Persival GmbH) + + [![avatar](../doc/img/avatar/JuergenWille.png) Jürgen Wille](https://github.com/FM-juergenW) (FrontMod GmbH) + ### Repositories - - #### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) - - Enter a short description of the model. - What is the purpose of the model? - What is the general modeling approach? - What inputs does the model need and what outputs does it generate? - - < Eye-catcher Image > - - - - more info: [click here](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) -
- -- #### [sl-1-1-reflection-based-radar-object-model](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model) +- #### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) + + > Enter a short description of the model. + > What is the purpose of the model? + > What is the general modeling approach? + > What inputs does the model need and what outputs does it generate? + > + > < Eye-catcher Image > + > + > + > + > more info: [click here](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) +
- - - This model is a Reflection Based Radar Model based on the [Modular OSMP Framework](https://gitlab.com/tuda-fzd/perception-sensor-modeling/modular-osmp-framework) by FZD. - It is a highly parameterizable sensor system model including detection calculation and object tracking simulation. - The model received radar reflection data calculated in a simulation tool beforehand e.g. with ray tracing. - The model outputs are radar detections and detected moving objects. - - +- #### [sl-1-1-reflection-based-radar-object-model](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model) - more info: [click here](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model) + > + > + > This model is a Reflection Based Radar Model based on the [Modular OSMP Framework](https://gitlab.com/tuda-fzd/perception-sensor-modeling/modular-osmp-framework) by FZD. + > It is a highly parameterizable sensor system model including detection calculation and object tracking simulation. + > The model received radar reflection data calculated in a simulation tool beforehand e.g. with ray tracing. + > The model outputs are radar detections and detected moving objects. + > + > + > + > more info: [click here](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model)
- + - #### [sl-1-2-reflection-based-lidar-object-model](https://github.com/openMSL/sl-1-2-reflection-based-lidar-object-model) - The current version of the model is build on the enhancements to the Open Simulation Interface from the publicly funded SETLevel project. - It is therefore dependent on the non-standard [SL OSI](https://gitlab.setlevel.de/open/osi) and not [ASAM OSI](https://github.com/OpenSimulationInterface/open-simulation-interface). - - - - This is the FZD Reflection Based Lidar Model based on the FZD OSI Sensor Model Packaging Framework. - It is a highly parameterizable sensor system model including detection calculation and object tracking simulation. - The model gets lidar reflection calculated in a simulation tool beforehand e.g. with ray tracing. - The model outputs are lidar detections and detected moving objects.
- - - - more info: [click here](https://github.com/openMSL/sl-1-2-reflection-based-lidar-object-model) + > The current version of the model is build on the enhancements to the Open Simulation Interface from the publicly funded SETLevel project. + > It is therefore dependent on the non-standard [SL OSI](https://gitlab.setlevel.de/open/osi) and not [ASAM OSI](https://github.com/OpenSimulationInterface/open-simulation-interface). + > + > + > + > This is the FZD Reflection Based Lidar Model based on the FZD OSI Sensor Model Packaging Framework. + > It is a highly parameterizable sensor system model including detection calculation and object tracking simulation. + > The model gets lidar reflection calculated in a simulation tool beforehand e.g. with ray tracing. + > The model outputs are lidar detections and detected moving objects.
+ > + > + > + > more info: [click here](https://github.com/openMSL/sl-1-2-reflection-based-lidar-object-model)
- + - #### [sl-1-3-object-based-generic-perception-object-model](https://github.com/openMSL/sl-1-3-object-based-generic-perception-object-model) - - - This model is a highly parameterizable generic perception sensor and tracking model. - It can be parameterized as a Lidar or a Radar. - The model is based on object lists and all modeling is performed on object level. - It includes typical sensor artifacts like soft FoV transitions, different detection ranges for different targets, - occlusion effects depending on the sensor technology as well as simulation of tracking behavior. - The model output are object lists for OSI SenorData moving objects. - - The architecture of the model as well as the parameterization structure are designed to be as generic as possible + > + > + > This model is a highly parameterizable generic perception sensor and tracking model. It can be parameterized as a Lidar or a Radar. The model is based on object lists and all modeling is performed on object level. + > It includes typical sensor artifacts like soft FoV transitions, different detection ranges for different targets occlusion effects depending on the sensor technology as well as simulation of tracking behavior. + > The model output are object lists for OSI SenorData moving objects. + > + > The architecture of the model as well as the parameterization structure are designed to be as generic as possible to fit both radar and lidar sensors to utilize similarities in signal propagation and signal processing in both technologies. - This way, the model can be parameterized to model different kinds of lidar and radar sensors. - To give an example: You can set an irradiation pattern for the modeled sensor. - Depending on the sensor technology this can either be an antenna gain pattern for radar or a beam pattern for lidar. - - - - more info: [click here](https://github.com/openMSL/sl-1-3-object-based-generic-perception-object-model) + > This way, the model can be parameterized to model different kinds of lidar and radar sensors. To give an example: You can set an irradiation pattern for the modeled sensor. Depending on the sensor technology this can either be an antenna gain pattern for radar or a beam pattern for lidar. + > + > + > + > more info: [click here](https://github.com/openMSL/sl-1-3-object-based-generic-perception-object-model)
- #### [sl-1-4-object-based-camera-object-model](https://github.com/openMSL/sl-1-4-object-based-camera-object-model) - This model is a parameterizable object based video perception sensor and tracking model using the interface OSI. - The model was developed in the project SetLevel by Bosch. The model should simulate some basic video typical effects in a phenomenological way. + > This model is a parameterizable object based video perception sensor and tracking model using the interface OSI. The model was developed in the project SetLevel by Bosch. The model should simulate some basic video typical effects in a phenomenological way. The "object based camera object model" is based on object lists and all modeling is performed on object level. The model output are object lists for OSI SenorData moving and stationary objects. - - The outer layer of the model is the OSI Sensor Model Packaging (OSMP). - It specifies ways in which models using the Open Simulation Interface (OSI) are to be packaged for their use in simulation environments using FMI 2.0. - For more detailed information see the official documentation. - - - - more info: [click here](https://github.com/openMSL/sl-1-4-object-based-camera-object-model) + > + > The outer layer of the model is the OSI Sensor Model Packaging (OSMP). It specifies ways in which models using the Open Simulation Interface (OSI) are to be packaged for their use in simulation environments using FMI 2.0. + > For more detailed information see the official documentation. + > + > + > + > more info: [click here](https://github.com/openMSL/sl-1-4-object-based-camera-object-model)
- -- #### [sl-1-5-sensor-model-testing](https://github.com/openMSL/sl-1-5-sensor-model-testing) - This repository contains FMUs for automated testing, verification and validation for open-source perception sensor models. +- #### [sl-1-5-sensor-model-testing](https://github.com/openMSL/sl-1-5-sensor-model-testing) - Further information can be found in [the repository](https://github.com/openMSL/sl-1-5-sensor-model-testing), the READMEs of the respective folders, and the documentation below: - - FMU collection for sensor model testing: - - [OSI Check FMU](src/osi-field-checker/)
- This FMU checks if fields are missing in a received SensorData. - - [SRMD Validator](src/srmd-validator/)
- This python script looks for SRMD files and validates them.
- -
+ > This repository contains FMUs for automated testing, verification and validation for open-source perception sensor models. + > + >Further information can be found in [the repository](https://github.com/openMSL/sl-1-5-sensor-model-testing), the READMEs of the respective folders, and the documentation below: + > + > - FMU collection for sensor model testing: + > - [OSI Check FMU](src/osi-field-checker/)
+ > This FMU checks if fields are missing in a received SensorData. + > - [SRMD Validator](src/srmd-validator/)
+ > This python script looks for SRMD files and validates them.
+
---- +---
## SL2 - Traffic Participant Models @@ -160,14 +172,14 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel - TBD - TBD - TBD - + ### Repositories - #### [sl-2-0-traffic-participant-model-repository-template](https://github.com/openMSL/sl-2-0-traffic-participant-model-repository-template)
---- +---
## SL3 - Scenario Data @@ -182,23 +194,25 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel - TBD (BMW AG) - TBD - TBD - + ### Repositories - #### [sl-3-1-osc-alks-scenarios](https://github.com/asam-oss/OSC-ALKS-scenarios) - The here provided 15 concrete parametrized test scenarios are derived from the 6 subject areas analogous to Annex 5, Chapter 4.1-4.6 as an initial attempt to clarify the described set of functional scenarios. - - Each concrete scenario is complemented by a variation file to form a logical scenario, which then represents a set of concrete scenarios. By applying the parameter variation defined in the variation files to the parameters in the concrete scenarios, multiple concrete scenarios can be derived or generated to cover the required scenario space. - - The focus of the here provided scenarios is on securing the planning aspects of an "Automated Lane Keeping System". By extending the scenarios with environmental conditions (e.g. light, rain or wind) or references to e.g. 3D models, aspects of sensor and actuator technology could also be simulated and validated. - - more info: [click here](https://github.com/asam-oss/OSC-ALKS-scenarios) + > The here provided 15 concrete parametrized test scenarios are derived from the 6 subject areas analogous to Annex 5, Chapter 4.1-4.6 as an initial attempt to clarify the described set of functional scenarios. + > + > Each concrete scenario is complemented by a variation file to form a logical scenario, which then represents a set of concrete scenarios. + > By applying the parameter variation defined in the variation files to the parameters in the concrete scenarios, multiple concrete scenarios can be derived or generated to cover the required scenario space. + > + > The focus of the here provided scenarios is on securing the planning aspects of an "Automated Lane Keeping System". + > By extending the scenarios with environmental conditions (e.g. light, rain or wind) or references to e.g. 3D models, aspects of sensor and actuator technology could also be simulated and validated. + > + > more info: [click here](https://github.com/asam-oss/OSC-ALKS-scenarios)

---- +---
## SL4 - Static Environment Data @@ -213,19 +227,18 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel - TBD (BMW AG) - TBD - TBD - + ### Repositories - In discussion
---- +---
## SL5 - Tooling - This sub-library contains various tools to import, export, analyze and visualize co-simulation data. Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hello@envited.market) @@ -235,60 +248,52 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel - TBD (Persival GmbH) - TBD - TBD - + ### Repositories - #### [sl-5-1-srmd-validator](https://github.com/openMSL/sl-5-1-srmd-validator) - This python code is meant to be used in a CI pipeline, e.g. a GitHub Action. -It looks for SRMD files in the root directory of a repository, this repo is cloned into. -The found SRMD files are validated against the SRMD schema from [SSPTraceability](https://github.com/PMSFIT/SSPTraceability/). - - more info: [click here](https://github.com/openMSL/sl-5-1-srmd-validator) + > This python code is meant to be used in a CI pipeline, e.g. a GitHub Action. It looks for SRMD files in the root directory of a repository, this repo is cloned into. The found SRMD files are validated against the SRMD schema from [SSPTraceability](https://github.com/PMSFIT/SSPTraceability/). + > + >more info: [click here](https://github.com/openMSL/sl-5-1-srmd-validator)
- + - #### [sl-5-2-osi-field-checker](https://github.com/openMSL/sl-5-2-osi-field-checker) - This FMU checks if fields are missing in a received SensorData. - It is meant to be used in a co-simulation connected to the output of the model under test. - It will output missing osi fields in the format of GitHub annotations, so that it can be used directly in a GitHub CI pipeline. - The image below shows an example of a failed pipeline due to missing OSI fields in the SensorData. - - OSI Field Checker in CI Pipeline - - more info: [click here](https://github.com/openMSL/sl-5-2-osi-field-checker) + > This FMU checks if fields are missing in a received SensorData. It is meant to be used in a co-simulation connected to the output of the model under test. It will output missing osi fields in the format of GitHub annotations, so that it can be used directly in a GitHub CI pipeline. + > The image below shows an example of a failed pipeline due to missing OSI fields in the SensorData. + > + > OSI Field Checker in CI Pipeline + > + >more info: [click here](https://github.com/openMSL/sl-5-2-osi-field-checker)
- #### [sl-5-3-osmp-network-proxy](https://github.com/openMSL/sl-5-3-osmp-network-proxy) - This Network Proxy FMU can receive SensorView and SensorData via TCP/IP using ZeroMQ. The received data is directly given to the FMU output. The proxy can also send SensorView oder SensorData received as FMU input via TCP/IP to a given IP address and port. + > This Network Proxy FMU can receive SensorView and SensorData via TCP/IP using ZeroMQ. The received data is directly given to the FMU output. The proxy can also send SensorView oder SensorData received as FMU input via TCP/IP to a given IP address and port. + > + > more info: [click here](https://github.com/openMSL/sl-5-3-osmp-network-proxy) +
- more info: [click here](https://github.com/openMSL/sl-5-3-osmp-network-proxy) -
- - #### [sl-5-4-standalone-osi-trace-file-player](https://github.com/openMSL/sl-5-4-standalone-osi-trace-file-player) - This mini application can read a binary ASAM OSI trace file (SensorData or SensorView) and send it step by step via TCP using ZeroMQ. - - more info: [click here](https://github.com/openMSL/sl-5-4-standalone-osi-trace-file-player) + > This mini application can read a binary ASAM OSI trace file (SensorData or SensorView) and send it step by step via TCP using ZeroMQ. + > + > more info: [click here](https://github.com/openMSL/sl-5-4-standalone-osi-trace-file-player)
- + - #### [sl-5-5-osi-trace-file-player](https://github.com/openMSL/sl-5-5-osi-trace-file-player) - This [FMU](https://fmi-standard.org/) is able to play binary OSI trace files. - The folder containing the trace files has to be passed as FMI parameter _trace_path_. - The trace file player is build according to the [ASAM Open simulation Interface (OSI)](https://github.com/OpenSimulationInterface/open-simulation-interface) and the [OSI Sensor Model Packaging (OSMP)](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) examples. - - more info: [click here](https://github.com/openMSL/sl-5-5-osi-trace-file-player) + > This [FMU](https://fmi-standard.org/) is able to play binary OSI trace files. The folder containing the trace files has to be passed as FMI parameter *trace_path*. + > The trace file player is build according to the [ASAM Open simulation Interface (OSI)](https://github.com/OpenSimulationInterface/open-simulation-interface) and the [OSI Sensor Model Packaging (OSMP)](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) examples. + > + > more info: [click here](https://github.com/openMSL/sl-5-5-osi-trace-file-player)
- -- #### [sl-5-6-osi-trace-file-writer](https://github.com/openMSL/sl-5-6-osi-trace-file-writer) - This [FMU](https://fmi-standard.org/) is able to write binary OSI SensorData trace files. - The folder the trace files shall be written to has to be passed as FMI parameter _trace_path_. - The trace file writer is build according to - the [ASAM Open simulation Interface (OSI)](https://github.com/OpenSimulationInterface/open-simulation-interface) and - the [OSI Sensor Model Packaging (OSMP)](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) examples. +- #### [sl-5-6-osi-trace-file-writer](https://github.com/openMSL/sl-5-6-osi-trace-file-writer) - more info: [click here](https://github.com/openMSL/sl-5-6-osi-trace-file-writer) + > This [FMU](https://fmi-standard.org/) is able to write binary OSI SensorData trace files. The folder the trace files shall be written to has to be passed as FMI parameter *trace_path*. + > The trace file writer is build according to the [ASAM Open simulation Interface (OSI)](https://github.com/OpenSimulationInterface/open-simulation-interface) and the [OSI Sensor Model Packaging (OSMP)](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) examples. + > + > more info: [click here](https://github.com/openMSL/sl-5-6-osi-trace-file-writer)
From 0980b783b5fafafc32f912c6bccc1da602761e75 Mon Sep 17 00:00:00 2001 From: Hadi Adineh Date: Tue, 19 Sep 2023 11:43:32 +0200 Subject: [PATCH 3/8] sl-1-5 removed from readme file Signed-off-by: Hadi Adineh --- profile/README.md | 14 -------------- 1 file changed, 14 deletions(-) diff --git a/profile/README.md b/profile/README.md index 0536242..7cd92c9 100644 --- a/profile/README.md +++ b/profile/README.md @@ -30,7 +30,6 @@ We are looking forward to welcome you as member of our community! - [sl-1-2-reflection-based-lidar-object-model](#sl-1-2-reflection-based-lidar-object-model) - [sl-1-3-object-based-generic-perception-object-model](#sl-1-3-object-based-generic-perception-object-model) - [sl-1-4-object-based-camera-object-model](#sl-1-4-object-based-camera-object-model) - - [sl-1-5-sensor-model-testing](#sl-1-5-sensor-model-testing) - [SL2 - Traffic Participant Models](#sl2---traffic-participant-models) - [sl-2-0-traffic-participant-model-repository-template](#sl-2-0-traffic-participant-model-repository-template) @@ -145,19 +144,6 @@ Initiated: 2022-07-25 > more info: [click here](https://github.com/openMSL/sl-1-4-object-based-camera-object-model)
-- #### [sl-1-5-sensor-model-testing](https://github.com/openMSL/sl-1-5-sensor-model-testing) - - > This repository contains FMUs for automated testing, verification and validation for open-source perception sensor models. - > - >Further information can be found in [the repository](https://github.com/openMSL/sl-1-5-sensor-model-testing), the READMEs of the respective folders, and the documentation below: - > - > - FMU collection for sensor model testing: - > - [OSI Check FMU](src/osi-field-checker/)
- > This FMU checks if fields are missing in a received SensorData. - > - [SRMD Validator](src/srmd-validator/)
- > This python script looks for SRMD files and validates them.
-
- ---
From fda74d6214a24df63cb84172c11d2edf4c190dc5 Mon Sep 17 00:00:00 2001 From: ClemensLinnhoff Date: Wed, 20 Sep 2023 14:31:15 +0200 Subject: [PATCH 4/8] Test headline link on sensor model template Signed-off-by: ClemensLinnhoff --- profile/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/profile/README.md b/profile/README.md index 7cd92c9..28d0542 100644 --- a/profile/README.md +++ b/profile/README.md @@ -69,7 +69,7 @@ Initiated: 2022-07-25 ### Repositories -- #### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) +#### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) > Enter a short description of the model. > What is the purpose of the model? From 7653385db5313b380c5ee7549a127fce2b27c20b Mon Sep 17 00:00:00 2001 From: ClemensLinnhoff Date: Wed, 20 Sep 2023 14:38:00 +0200 Subject: [PATCH 5/8] Test custom named anchor Signed-off-by: ClemensLinnhoff --- profile/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/profile/README.md b/profile/README.md index 28d0542..86aa57c 100644 --- a/profile/README.md +++ b/profile/README.md @@ -69,7 +69,7 @@ Initiated: 2022-07-25 ### Repositories -#### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) +#### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) {#sl-1-0-sensor-model-repository-template} > Enter a short description of the model. > What is the purpose of the model? From c1a9b2a78017fbbda218ebc7efdbc2f23a897d40 Mon Sep 17 00:00:00 2001 From: ClemensLinnhoff Date: Wed, 20 Sep 2023 14:41:40 +0200 Subject: [PATCH 6/8] Reverse named anchor Signed-off-by: ClemensLinnhoff --- profile/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/profile/README.md b/profile/README.md index 86aa57c..28d0542 100644 --- a/profile/README.md +++ b/profile/README.md @@ -69,7 +69,7 @@ Initiated: 2022-07-25 ### Repositories -#### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) {#sl-1-0-sensor-model-repository-template} +#### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) > Enter a short description of the model. > What is the purpose of the model? From 97cae6a84a30af9fc66594ef1799f0e648225ef0 Mon Sep 17 00:00:00 2001 From: Hadi Adineh Date: Thu, 5 Oct 2023 14:23:20 +0200 Subject: [PATCH 7/8] removed images, heading lists and white spaces (#27) Signed-off-by: Hadi Adineh --- profile/README.md | 77 +++++++++++------------------------------------ 1 file changed, 18 insertions(+), 59 deletions(-) diff --git a/profile/README.md b/profile/README.md index 28d0542..6c4619b 100644 --- a/profile/README.md +++ b/profile/README.md @@ -47,10 +47,7 @@ We are looking forward to welcome you as member of our community! - [sl-5-5-osi-trace-file-player](#sl-5-5-osi-trace-file-player) - [sl-5-6-osi-trace-file-writer](#sl-5-6-osi-trace-file-writer) -
- --- -
## SL1 - Perception Sensor Models @@ -69,83 +66,59 @@ Initiated: 2022-07-25 ### Repositories -#### [sl-1-0-sensor-model-repository-template](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) +#### sl-1-0-sensor-model-repository-template > Enter a short description of the model. > What is the purpose of the model? > What is the general modeling approach? > What inputs does the model need and what outputs does it generate? > - > < Eye-catcher Image > - > - > - > > more info: [click here](https://github.com/openMSL/sl-1-0-sensor-model-repository-template) -
-- #### [sl-1-1-reflection-based-radar-object-model](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model) +#### sl-1-1-reflection-based-radar-object-model - > - > > This model is a Reflection Based Radar Model based on the [Modular OSMP Framework](https://gitlab.com/tuda-fzd/perception-sensor-modeling/modular-osmp-framework) by FZD. > It is a highly parameterizable sensor system model including detection calculation and object tracking simulation. > The model received radar reflection data calculated in a simulation tool beforehand e.g. with ray tracing. > The model outputs are radar detections and detected moving objects. > - > - > > more info: [click here](https://github.com/openMSL/sl-1-1-reflection-based-radar-object-model) -
-- #### [sl-1-2-reflection-based-lidar-object-model](https://github.com/openMSL/sl-1-2-reflection-based-lidar-object-model) +#### sl-1-2-reflection-based-lidar-object-model > The current version of the model is build on the enhancements to the Open Simulation Interface from the publicly funded SETLevel project. > It is therefore dependent on the non-standard [SL OSI](https://gitlab.setlevel.de/open/osi) and not [ASAM OSI](https://github.com/OpenSimulationInterface/open-simulation-interface). > - > - > > This is the FZD Reflection Based Lidar Model based on the FZD OSI Sensor Model Packaging Framework. > It is a highly parameterizable sensor system model including detection calculation and object tracking simulation. > The model gets lidar reflection calculated in a simulation tool beforehand e.g. with ray tracing. > The model outputs are lidar detections and detected moving objects.
> - > - > > more info: [click here](https://github.com/openMSL/sl-1-2-reflection-based-lidar-object-model) -
-- #### [sl-1-3-object-based-generic-perception-object-model](https://github.com/openMSL/sl-1-3-object-based-generic-perception-object-model) +#### sl-1-3-object-based-generic-perception-object-model - > - > > This model is a highly parameterizable generic perception sensor and tracking model. It can be parameterized as a Lidar or a Radar. The model is based on object lists and all modeling is performed on object level. > It includes typical sensor artifacts like soft FoV transitions, different detection ranges for different targets occlusion effects depending on the sensor technology as well as simulation of tracking behavior. > The model output are object lists for OSI SenorData moving objects. > > The architecture of the model as well as the parameterization structure are designed to be as generic as possible - to fit both radar and lidar sensors to utilize similarities in signal propagation and signal processing in both technologies. + > to fit both radar and lidar sensors to utilize similarities in signal propagation and signal processing in both technologies. > This way, the model can be parameterized to model different kinds of lidar and radar sensors. To give an example: You can set an irradiation pattern for the modeled sensor. Depending on the sensor technology this can either be an antenna gain pattern for radar or a beam pattern for lidar. > - > - > > more info: [click here](https://github.com/openMSL/sl-1-3-object-based-generic-perception-object-model) -
-- #### [sl-1-4-object-based-camera-object-model](https://github.com/openMSL/sl-1-4-object-based-camera-object-model) +#### sl-1-4-object-based-camera-object-model > This model is a parameterizable object based video perception sensor and tracking model using the interface OSI. The model was developed in the project SetLevel by Bosch. The model should simulate some basic video typical effects in a phenomenological way. - The "object based camera object model" is based on object lists and all modeling is performed on object level. The model output are object lists for OSI SenorData moving and stationary objects. + > The "object based camera object model" is based on object lists and all modeling is performed on object level. The model output are object lists for OSI SenorData moving and stationary objects. > > The outer layer of the model is the OSI Sensor Model Packaging (OSMP). It specifies ways in which models using the Open Simulation Interface (OSI) are to be packaged for their use in simulation environments using FMI 2.0. > For more detailed information see the official documentation. > - > - > > more info: [click here](https://github.com/openMSL/sl-1-4-object-based-camera-object-model) -
--- -
## SL2 - Traffic Participant Models @@ -161,12 +134,13 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel ### Repositories -- #### [sl-2-0-traffic-participant-model-repository-template](https://github.com/openMSL/sl-2-0-traffic-participant-model-repository-template) +#### sl-2-0-traffic-participant-model-repository-template -
+ > Under development + > + > more info: [click here](https://github.com/openMSL/sl-2-0-traffic-participant-model-repository-template) --- -
## SL3 - Scenario Data @@ -183,7 +157,7 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel ### Repositories -- #### [sl-3-1-osc-alks-scenarios](https://github.com/asam-oss/OSC-ALKS-scenarios) +#### sl-3-1-osc-alks-scenarios > The here provided 15 concrete parametrized test scenarios are derived from the 6 subject areas analogous to Annex 5, Chapter 4.1-4.6 as an initial attempt to clarify the described set of functional scenarios. > @@ -194,12 +168,8 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel > By extending the scenarios with environmental conditions (e.g. light, rain or wind) or references to e.g. 3D models, aspects of sensor and actuator technology could also be simulated and validated. > > more info: [click here](https://github.com/asam-oss/OSC-ALKS-scenarios) -
- -
--- -
## SL4 - Static Environment Data @@ -218,10 +188,7 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel - In discussion -
- --- -
## SL5 - Tooling @@ -237,49 +204,41 @@ Initiated: Call for participation. Get engaged [hello@envited.market](mailto:hel ### Repositories -- #### [sl-5-1-srmd-validator](https://github.com/openMSL/sl-5-1-srmd-validator) +#### sl-5-1-srmd-validator > This python code is meant to be used in a CI pipeline, e.g. a GitHub Action. It looks for SRMD files in the root directory of a repository, this repo is cloned into. The found SRMD files are validated against the SRMD schema from [SSPTraceability](https://github.com/PMSFIT/SSPTraceability/). > >more info: [click here](https://github.com/openMSL/sl-5-1-srmd-validator) -
-- #### [sl-5-2-osi-field-checker](https://github.com/openMSL/sl-5-2-osi-field-checker) +#### sl-5-2-osi-field-checker > This FMU checks if fields are missing in a received SensorData. It is meant to be used in a co-simulation connected to the output of the model under test. It will output missing osi fields in the format of GitHub annotations, so that it can be used directly in a GitHub CI pipeline. > The image below shows an example of a failed pipeline due to missing OSI fields in the SensorData. > - > OSI Field Checker in CI Pipeline - > >more info: [click here](https://github.com/openMSL/sl-5-2-osi-field-checker) -
-- #### [sl-5-3-osmp-network-proxy](https://github.com/openMSL/sl-5-3-osmp-network-proxy) +#### sl-5-3-osmp-network-proxy > This Network Proxy FMU can receive SensorView and SensorData via TCP/IP using ZeroMQ. The received data is directly given to the FMU output. The proxy can also send SensorView oder SensorData received as FMU input via TCP/IP to a given IP address and port. > > more info: [click here](https://github.com/openMSL/sl-5-3-osmp-network-proxy) -
-- #### [sl-5-4-standalone-osi-trace-file-player](https://github.com/openMSL/sl-5-4-standalone-osi-trace-file-player) +#### sl-5-4-standalone-osi-trace-file-player > This mini application can read a binary ASAM OSI trace file (SensorData or SensorView) and send it step by step via TCP using ZeroMQ. > > more info: [click here](https://github.com/openMSL/sl-5-4-standalone-osi-trace-file-player) -
-- #### [sl-5-5-osi-trace-file-player](https://github.com/openMSL/sl-5-5-osi-trace-file-player) +#### sl-5-5-osi-trace-file-player > This [FMU](https://fmi-standard.org/) is able to play binary OSI trace files. The folder containing the trace files has to be passed as FMI parameter *trace_path*. > The trace file player is build according to the [ASAM Open simulation Interface (OSI)](https://github.com/OpenSimulationInterface/open-simulation-interface) and the [OSI Sensor Model Packaging (OSMP)](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) examples. > > more info: [click here](https://github.com/openMSL/sl-5-5-osi-trace-file-player) -
-- #### [sl-5-6-osi-trace-file-writer](https://github.com/openMSL/sl-5-6-osi-trace-file-writer) +#### sl-5-6-osi-trace-file-writer > This [FMU](https://fmi-standard.org/) is able to write binary OSI SensorData trace files. The folder the trace files shall be written to has to be passed as FMI parameter *trace_path*. > The trace file writer is build according to the [ASAM Open simulation Interface (OSI)](https://github.com/OpenSimulationInterface/open-simulation-interface) and the [OSI Sensor Model Packaging (OSMP)](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) examples. > > more info: [click here](https://github.com/openMSL/sl-5-6-osi-trace-file-writer) -
From d875d1413ae937711b031456886e3a5ad19c2edc Mon Sep 17 00:00:00 2001 From: Hadi Adineh Date: Thu, 5 Oct 2023 14:33:43 +0200 Subject: [PATCH 8/8] removed images, heading lists and white spaces (#27) Signed-off-by: Hadi Adineh --- profile/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/profile/README.md b/profile/README.md index 6c4619b..cfc5347 100644 --- a/profile/README.md +++ b/profile/README.md @@ -51,7 +51,7 @@ We are looking forward to welcome you as member of our community! ## SL1 - Perception Sensor Models -> This sub-library is a collection of [OSI](https://github.com/OpenSimulationInterface/open-simulation-interface) compliant sensor models according to the [OSMP](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) specification including a template repository +This sub-library is a collection of [OSI](https://github.com/OpenSimulationInterface/open-simulation-interface) compliant sensor models according to the [OSMP](https://github.com/OpenSimulationInterface/osi-sensor-model-packaging) specification including a template repository demonstrating the [Credible Simulation Process](https://setlevel.de/assets/forschungsergebnisse/Credible-Simulation-Process-v1.0.pdf) by running full scale [SSP](https://ssp-standard.org/) based co-simulations in the CI pipeline. Initiated: 2022-07-25