GETTING STARTED
Indepth theory
Key concepts
9min
do i need to read this page? if you have already gone through the quickstart docid\ b6im5u kiwn9ukybzjx6d page then you usually don't have to read this page this is page covers more background information project project is the top most concept when interfacing with the kognic platform it is possible to have multiple ongoing projects, and they act as a container for other kognic resources project setup is usually performed by the kognic professional services team during the guideline agreement process (gap) of a new client engagement within the kognic apis, projects are identified using an external identifier batch input batches allow grouping of input data into smaller batches within a project by default, every project has at least one input batch ongoing projects can be benefited from having multiple batches in two ways group input data that are collected during a certain time interval perform guideline or task definition changes without the need for retroactive changes batch status status description pending batch has been created but is still not fully configured by kognic either during project setup or requested changes open batch is open for new inputs ready batch has been published and no longer open for new inputs in progress kognic has started annotation of inputs within the batch completed annotations has been completed request during gap, projects can have several annotation types as the end goal for example, a project consisting of images can be assigned for both lane detection and object annotation within kognic, a request represents a specific annotation goal for a given input we divide big and complex projects into several independent annotation types this makes it possible to reduce the cognitive load on the annotators more annotators can work on the same data in parallel simplify user interfaces all of these contribute to a high level of quality while also reducing the total time needed for producing an annotation guideline in order to produce annotations, one needs to know what to annotate and how this type of information is found in guideline a guideline defines what specific object to mark (e g vehicles and pedestrians), as well as how (e g bounding box) a guideline also includes detailed information about how to interpret the data, e g what it means by a vehicle is "heavily occluded" task definition task definition describes what should/could be annotated how many object types? bounding box, semantic segmentation or lines/splines for each object type? what are the properties for each object type? task definitions are json files that the kognic professional services team generates from the guideline the task definition is then used by the kognic app to construct the appropriate drawing tool in other words, task definition can be understood as the machine readable quivalent of a guideline scene before setting up any annotation task, the raw data needs to be correctly uploaded to the kognic platform the scene specifies how data from different sources are combined together resources are images and point clouds, as well as metadata and calibrations (define sensors' properties) we support different types of setup, for example image(s) from a (multiple) camera(s) image(s) from camera(s) combined with lidar point clouds another concept related to scene is frame a frame is a discrete moment of a scene in time scenes can be either single frame or sequence (multiple frames) sequence should be used when temporal information is important for producing the annotation scene types type description cameras a single frame consisting of image(s) from 1 9 cameras lidarsandcameras a single frame consisting of 1 20 lidar point clouds and image(s) from 1 9 cameras camerasseq a sequence of frames, each frame consisting of image(s) from 1 9 cameras lidarsandcamerasseq a sequence of frames, each frame consisting of 1 20 lidar point clouds and image(s) from 1 9 cameras aggregatedlidarsandcamerasseq a sequence of frames, each frame consisting of 1 20 lidar point clouds and image(s) from 1 9 cameras however, point clouds are aggregated over time, producing a unified point cloud input once a scene has been uploaded to the kognic platform, one can create annotation tasks as inputs where each input is associated to a request differenciate input from scene enables efficient reuse of the uploaded data for instance, multiple inputs can be created from the same scene enabling different kinds of annotation setups note that one can create an input simultaneously when creating a scene by providing the project/batch that it should be associated to, see examples in working with scenes & inputs docid\ utyshj bng1v85n2e6 eh annotation an annotation is produced when inputs are successfully annotated in a request annotations are provided by kognic io api as json objects in openlabel format docid\ awhhsrojm96zowliryw7d more information on how to download these annotations along with some examples of how to work with them is available in the download annotations docid 5j mcgjyo5xffxqkkrqgo chapter apart from kognic io api, kognic also provides a library called kognic openlabel docid\ s0nqg6db56u4sihk3jtlx , which makes it easy to parse and work with the openlabel json objects