Skip to content

API documentation of the mongodb provider

The module that contains all the necessary logic for communication with the MongoDb storage providers.

MongodbCore

Bases: StorageCore

Base class that creates the most important functions for the mongodb storage provider.

Source code in src/sqooler/storage_providers/mongodb.py
 35
 36
 37
 38
 39
 40
 41
 42
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
class MongodbCore(StorageCore):
    """
    Base class that creates the most important functions for the mongodb storage provider.
    """

    def __init__(
        self, login_dict: MongodbLoginInformation, name: str, is_active: bool = True
    ) -> None:
        """
        Set up the neccessary keys and create the client through which all the connections will run.

        Args:
            login_dict: The login dict that contains the neccessary
                        information to connect to the mongodb
            name: The name of the storage provider
            is_active: Is the storage provider active.


        Raises:
            ValidationError: If the login_dict is not valid
        """
        super().__init__(name, is_active)
        mongodb_username = login_dict.mongodb_username
        mongodb_password = login_dict.mongodb_password
        mongodb_database_url = login_dict.mongodb_database_url

        uri = f"mongodb+srv://{mongodb_username}:{mongodb_password}@{mongodb_database_url}"
        uri = uri + "/?retryWrites=true&w=majority"
        # Create a new client and connect to the server
        self.client: MongoClient = MongoClient(uri)

        # Send a ping to confirm a successful connection
        self.client.admin.command("ping")

    @validate_active
    def upload(self, content_dict: dict, storage_path: str, job_id: str) -> None:
        """
        Upload the file to the storage

        content_dict: the content that should be uploaded onto the mongodb base
        storage_path: the access path towards the mongodb collection
        job_id: the id of the file we are about to create
        """

        _, collection = self._get_database_and_collection(storage_path)

        content_dict["_id"] = ObjectId(job_id)

        try:
            collection.insert_one(content_dict)
        except DuplicateKeyError as err:
            raise FileExistsError(
                f"The file with the id {job_id} already exists in the collection {storage_path}."
            ) from err
        # remove the id from the content dict for further use
        content_dict.pop("_id", None)

    @validate_active
    def get(self, storage_path: str, job_id: str) -> dict:
        """
        Get the file content from the storage

        Args:
            storage_path: the path towards the file, excluding the filename / id
            job_id: the id of the file we are about to look up

        Returns:
            The content of the file
        """
        try:
            document_to_find = {"_id": ObjectId(job_id)}
        except InvalidId as err:
            raise FileNotFoundError(
                f"The job_id {job_id} is not valid. Please check the job_id."
            ) from err

        document_to_find = {"_id": ObjectId(job_id)}

        _, collection = self._get_database_and_collection(storage_path)

        result_found = collection.find_one(document_to_find)

        if not result_found:
            raise FileNotFoundError(
                f"Could not find a file under {storage_path} with the id {job_id}."
            )

        # remove the id from the result dict for further use
        result_found.pop("_id", None)
        return result_found

    @validate_active
    def update(self, content_dict: dict, storage_path: str, job_id: str) -> None:
        """
        Update the file content. It replaces the old content with the new content.


        Args:
            content_dict: The dictionary containing the new content of the file
            storage_path: The path to the file
            job_id: The id of the job

        Returns:
            None

        Raises:
            FileNotFoundError: If the file is not found
        """

        _, collection = self._get_database_and_collection(storage_path)

        filter_dict = {"_id": ObjectId(job_id)}
        result = collection.replace_one(filter_dict, content_dict)

        if result.matched_count == 0:
            raise FileNotFoundError(f"Could not update file under {storage_path}")

    @validate_active
    def move(self, start_path: str, final_path: str, job_id: str) -> None:
        """
        Move the file from start_path to final_path

        start_path: the path where the file is currently stored, but excluding the file name
        final_path: the path where the file should be stored, but excluding the file name
        job_id: the name of the file. Is a json file

        Returns:
            None
        """

        # delete the old file
        _, collection = self._get_database_and_collection(start_path)

        document_to_find = {"_id": ObjectId(job_id)}
        result_found = collection.find_one(document_to_find)

        collection.delete_one(document_to_find)

        # add the document to the new collection
        _, collection = self._get_database_and_collection(final_path)

        collection.insert_one(result_found)

    @validate_active
    def delete(self, storage_path: str, job_id: str) -> None:
        """
        Remove the file from the mongodb database

        Args:
            storage_path: the path where the file is currently stored, but excluding the file name
            job_id: the name of the file

        Returns:
            None
        """
        _, collection = self._get_database_and_collection(storage_path)

        try:
            document_to_find = {"_id": ObjectId(job_id)}
        except InvalidId as err:
            raise FileNotFoundError(
                f"The job_id {job_id} is not valid. Please check the job_id."
            ) from err
        result = collection.delete_one(document_to_find)
        if result.deleted_count == 0:
            raise FileNotFoundError(
                f"Could not find a file under {storage_path} with the id {job_id}."
            )

    def _get_database_and_collection(
        self, storage_path: str
    ) -> tuple[Database, Collection]:
        """
        Get the database and the collection on which we work.

        Args:
            storage_path: the path where the file is currently stored, but excluding the file name

        Returns:
            The database and the collection on which we work
        """
        # strip the path from leading and trailing slashes
        storage_path = storage_path.strip("/")

        # get the database on which we work
        database = self.client[storage_path.split("/")[0]]

        # get the collection on which we work
        collection_name = ".".join(storage_path.split("/")[1:])
        collection = database[collection_name]
        return database, collection

__init__(login_dict, name, is_active=True)

Set up the neccessary keys and create the client through which all the connections will run.

Parameters:

Name Type Description Default
login_dict MongodbLoginInformation

The login dict that contains the neccessary information to connect to the mongodb

required
name str

The name of the storage provider

required
is_active bool

Is the storage provider active.

True

Raises:

Type Description
ValidationError

If the login_dict is not valid

Source code in src/sqooler/storage_providers/mongodb.py
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
def __init__(
    self, login_dict: MongodbLoginInformation, name: str, is_active: bool = True
) -> None:
    """
    Set up the neccessary keys and create the client through which all the connections will run.

    Args:
        login_dict: The login dict that contains the neccessary
                    information to connect to the mongodb
        name: The name of the storage provider
        is_active: Is the storage provider active.


    Raises:
        ValidationError: If the login_dict is not valid
    """
    super().__init__(name, is_active)
    mongodb_username = login_dict.mongodb_username
    mongodb_password = login_dict.mongodb_password
    mongodb_database_url = login_dict.mongodb_database_url

    uri = f"mongodb+srv://{mongodb_username}:{mongodb_password}@{mongodb_database_url}"
    uri = uri + "/?retryWrites=true&w=majority"
    # Create a new client and connect to the server
    self.client: MongoClient = MongoClient(uri)

    # Send a ping to confirm a successful connection
    self.client.admin.command("ping")

delete(storage_path, job_id)

Remove the file from the mongodb database

Parameters:

Name Type Description Default
storage_path str

the path where the file is currently stored, but excluding the file name

required
job_id str

the name of the file

required

Returns:

Type Description
None

None

Source code in src/sqooler/storage_providers/mongodb.py
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
@validate_active
def delete(self, storage_path: str, job_id: str) -> None:
    """
    Remove the file from the mongodb database

    Args:
        storage_path: the path where the file is currently stored, but excluding the file name
        job_id: the name of the file

    Returns:
        None
    """
    _, collection = self._get_database_and_collection(storage_path)

    try:
        document_to_find = {"_id": ObjectId(job_id)}
    except InvalidId as err:
        raise FileNotFoundError(
            f"The job_id {job_id} is not valid. Please check the job_id."
        ) from err
    result = collection.delete_one(document_to_find)
    if result.deleted_count == 0:
        raise FileNotFoundError(
            f"Could not find a file under {storage_path} with the id {job_id}."
        )

get(storage_path, job_id)

Get the file content from the storage

Parameters:

Name Type Description Default
storage_path str

the path towards the file, excluding the filename / id

required
job_id str

the id of the file we are about to look up

required

Returns:

Type Description
dict

The content of the file

Source code in src/sqooler/storage_providers/mongodb.py
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
@validate_active
def get(self, storage_path: str, job_id: str) -> dict:
    """
    Get the file content from the storage

    Args:
        storage_path: the path towards the file, excluding the filename / id
        job_id: the id of the file we are about to look up

    Returns:
        The content of the file
    """
    try:
        document_to_find = {"_id": ObjectId(job_id)}
    except InvalidId as err:
        raise FileNotFoundError(
            f"The job_id {job_id} is not valid. Please check the job_id."
        ) from err

    document_to_find = {"_id": ObjectId(job_id)}

    _, collection = self._get_database_and_collection(storage_path)

    result_found = collection.find_one(document_to_find)

    if not result_found:
        raise FileNotFoundError(
            f"Could not find a file under {storage_path} with the id {job_id}."
        )

    # remove the id from the result dict for further use
    result_found.pop("_id", None)
    return result_found

move(start_path, final_path, job_id)

Move the file from start_path to final_path

start_path: the path where the file is currently stored, but excluding the file name final_path: the path where the file should be stored, but excluding the file name job_id: the name of the file. Is a json file

Returns:

Type Description
None

None

Source code in src/sqooler/storage_providers/mongodb.py
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
@validate_active
def move(self, start_path: str, final_path: str, job_id: str) -> None:
    """
    Move the file from start_path to final_path

    start_path: the path where the file is currently stored, but excluding the file name
    final_path: the path where the file should be stored, but excluding the file name
    job_id: the name of the file. Is a json file

    Returns:
        None
    """

    # delete the old file
    _, collection = self._get_database_and_collection(start_path)

    document_to_find = {"_id": ObjectId(job_id)}
    result_found = collection.find_one(document_to_find)

    collection.delete_one(document_to_find)

    # add the document to the new collection
    _, collection = self._get_database_and_collection(final_path)

    collection.insert_one(result_found)

update(content_dict, storage_path, job_id)

Update the file content. It replaces the old content with the new content.

Parameters:

Name Type Description Default
content_dict dict

The dictionary containing the new content of the file

required
storage_path str

The path to the file

required
job_id str

The id of the job

required

Returns:

Type Description
None

None

Raises:

Type Description
FileNotFoundError

If the file is not found

Source code in src/sqooler/storage_providers/mongodb.py
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
@validate_active
def update(self, content_dict: dict, storage_path: str, job_id: str) -> None:
    """
    Update the file content. It replaces the old content with the new content.


    Args:
        content_dict: The dictionary containing the new content of the file
        storage_path: The path to the file
        job_id: The id of the job

    Returns:
        None

    Raises:
        FileNotFoundError: If the file is not found
    """

    _, collection = self._get_database_and_collection(storage_path)

    filter_dict = {"_id": ObjectId(job_id)}
    result = collection.replace_one(filter_dict, content_dict)

    if result.matched_count == 0:
        raise FileNotFoundError(f"Could not update file under {storage_path}")

upload(content_dict, storage_path, job_id)

Upload the file to the storage

content_dict: the content that should be uploaded onto the mongodb base storage_path: the access path towards the mongodb collection job_id: the id of the file we are about to create

Source code in src/sqooler/storage_providers/mongodb.py
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
@validate_active
def upload(self, content_dict: dict, storage_path: str, job_id: str) -> None:
    """
    Upload the file to the storage

    content_dict: the content that should be uploaded onto the mongodb base
    storage_path: the access path towards the mongodb collection
    job_id: the id of the file we are about to create
    """

    _, collection = self._get_database_and_collection(storage_path)

    content_dict["_id"] = ObjectId(job_id)

    try:
        collection.insert_one(content_dict)
    except DuplicateKeyError as err:
        raise FileExistsError(
            f"The file with the id {job_id} already exists in the collection {storage_path}."
        ) from err
    # remove the id from the content dict for further use
    content_dict.pop("_id", None)

MongodbProvider

Bases: MongodbProviderExtended

The access to the mongodb. This is the simplified version for people that are running devices.

Source code in src/sqooler/storage_providers/mongodb.py
756
757
758
759
760
761
762
763
764
765
class MongodbProvider(MongodbProviderExtended):
    """
    The access to the mongodb. This is the simplified version for people that are running devices.
    """

    def __init__(self, login_dict: MongodbLoginInformation) -> None:
        """
        Set up the neccessary keys and create the client through which all the connections will run.
        """
        super().__init__(login_dict, name="default", is_active=True)

__init__(login_dict)

Set up the neccessary keys and create the client through which all the connections will run.

Source code in src/sqooler/storage_providers/mongodb.py
761
762
763
764
765
def __init__(self, login_dict: MongodbLoginInformation) -> None:
    """
    Set up the neccessary keys and create the client through which all the connections will run.
    """
    super().__init__(login_dict, name="default", is_active=True)

MongodbProviderExtended

Bases: StorageProvider, MongodbCore

The access to the mongodb

Attributes:

Name Type Description
configs_path PathStr

The path to the folder where the configurations are stored

queue_path PathStr

The path to the folder where the jobs are stored

running_path PathStr

The path to the folder where the running jobs are stored

finished_path PathStr

The path to the folder where the finished jobs are stored

deleted_path PathStr

The path to the folder where the deleted jobs are stored

status_path PathStr

The path to the folder where the status is stored

results_path PathStr

The path to the folder where the results are stored

pks_path PathStr

The path to the folder where the public keys are stored

Source code in src/sqooler/storage_providers/mongodb.py
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
class MongodbProviderExtended(StorageProvider, MongodbCore):
    """
    The access to the mongodb

    Attributes:
        configs_path: The path to the folder where the configurations are stored
        queue_path: The path to the folder where the jobs are stored
        running_path: The path to the folder where the running jobs are stored
        finished_path: The path to the folder where the finished jobs are stored
        deleted_path: The path to the folder where the deleted jobs are stored
        status_path: The path to the folder where the status is stored
        results_path: The path to the folder where the results are stored
        pks_path: The path to the folder where the public keys are stored
    """

    configs_path: PathStr = "backends/configs"
    queue_path: PathStr = "jobs/queued"
    running_path: PathStr = "jobs/running"
    finished_path: PathStr = "jobs/finished"
    deleted_path: PathStr = "jobs/deleted"
    status_path: PathStr = "status"
    results_path: PathStr = "results"
    pks_path: PathStr = "backends/public_keys"

    def get_attribute_path(
        self,
        attribute_name: AttributePathStr,
        display_name: Optional[DisplayNameStr] = None,
        job_id: Optional[str] = None,
        username: Optional[str] = None,
    ) -> str:
        """
        Get the path to the results of the device.

        Args:
            display_name: The name of the backend
            attribute_name: The name of the attribute
            job_id: The job_id of the job
            username: The username of the user

        Returns:
            The path to the results of the device.
        """

        match attribute_name:
            case "configs":
                path = self.configs_path
            case "results":
                path = f"{self.results_path}/{display_name}"
            case "running":
                path = self.running_path
            case "status":
                path = f"{self.status_path}/{display_name}"
            case "queue":
                path = f"{self.queue_path}/{display_name}"
            case "deleted":
                path = self.deleted_path
            case "finished":
                path = f"{self.finished_path}/{display_name}"
            case "pks":
                path = self.pks_path
            case _:
                raise ValueError(f"The attribute name {attribute_name} is not valid.")
        return path

    def get_attribute_id(
        self,
        attribute_name: AttributeIdStr,
        job_id: str,
        display_name: Optional[DisplayNameStr] = None,
    ) -> str:
        """
        Get the path to the id of the device.

        Args:
            attribute_name: The name of the attribute
            job_id: The job_id of the job
            display_name: The name of the backend

        Returns:
            The path to the results of the device.
        """

        match attribute_name:
            case "configs":
                raise ValueError(f"The attribute name {attribute_name} is not valid.")
            case "job":
                _id = job_id
            case "results":
                _id = job_id
            case "status":
                _id = job_id
            case _:
                raise ValueError(f"The attribute name {attribute_name} is not valid.")
        return _id

    @validate_active
    def get_backends(self) -> list[DisplayNameStr]:
        """
        Get a list of all the backends that the provider offers.
        """

        # get the collection on which we work
        _, config_collection = self._get_database_and_collection(self.configs_path)

        # get all the documents in the collection configs and save the disply_name in a list
        backend_names: list[DisplayNameStr] = []
        for config_dict in config_collection.find():
            config_dict.pop("_id")
            expected_keys_for_jws = {"header", "payload", "signature"}
            if set(config_dict.keys()) == expected_keys_for_jws:
                backend_names.append(config_dict["payload"]["display_name"])
            else:
                backend_names.append(config_dict["display_name"])
        return backend_names

    def upload_config(
        self,
        config_dict: BackendConfigSchemaIn,
        display_name: DisplayNameStr,
        private_jwk: Optional[JWK] = None,
    ) -> None:
        """
        The function that uploads the spooler configuration to the storage.

        Args:
            config_dict: The dictionary containing the configuration
            display_name : The name of the backend
            private_jwk: The private JWK to sign the configuration with

        Returns:
            None
        """
        config_dict = self._verify_config(config_dict, display_name)

        # first we have to check if the device already exists in the database

        document_to_find = {"display_name": display_name}

        # get the collection on which we work
        _, collection = self._get_database_and_collection(self.configs_path)

        document_to_find = {"display_name": display_name}
        result_found = collection.find_one(document_to_find)
        if result_found:
            raise FileExistsError(
                f"The configuration for {display_name} already exists and should not be overwritten."
            )

        # now also look for signed configurations
        signed_document_to_find = {"payload.display_name": display_name}
        result_found = collection.find_one(signed_document_to_find)
        if result_found:
            raise FileExistsError(
                f"The configuration for {display_name} already exists and should not be overwritten."
            )

        upload_dict = self._format_config_dict(config_dict, private_jwk)
        config_id = uuid.uuid4().hex[:24]
        self.upload(upload_dict, self.configs_path, config_id)

    def update_config(
        self,
        config_dict: BackendConfigSchemaIn,
        display_name: DisplayNameStr,
        private_jwk: Optional[JWK] = None,
    ) -> None:
        """
        The function that updates the spooler configuration on the storage.

        Args:
            config_dict: The dictionary containing the configuration
            display_name : The name of the backend
            private_jwk: The private key of the backend

        Returns:
            None
        """

        config_dict = self._verify_config(config_dict, display_name)

        _, collection = self._get_database_and_collection(self.configs_path)

        # now make sure that we add the timezone as we open the file
        collection_with_tz = collection.with_options(
            codec_options=CodecOptions(tz_aware=True, tzinfo=timezone.utc)
        )
        # first we have to check if the device already exists in the database
        document_to_find = {"display_name": display_name}
        result_found = collection_with_tz.find_one(document_to_find)

        signed_document_to_find = {"payload.display_name": display_name}
        signed_backend_config_dict = collection_with_tz.find_one(
            signed_document_to_find
        )

        if result_found:
            old_config_jws = result_found
            job_id = result_found["_id"]
        elif signed_backend_config_dict:
            old_config_jws = signed_backend_config_dict
            job_id = signed_backend_config_dict["_id"]
            old_config_jws.pop("_id")
        else:
            raise FileNotFoundError(
                (
                    f"The config for {display_name} does not exist and should not be updated."
                    "Use the upload_config method instead."
                )
            )
        upload_dict = self._format_update_config(
            old_config_jws, config_dict, private_jwk
        )

        self.update(
            content_dict=upload_dict,
            storage_path=self.configs_path,
            job_id=job_id,
        )

    @validate_active
    def get_config(self, display_name: DisplayNameStr) -> BackendConfigSchemaIn:
        """
        The function that downloads the spooler configuration to the storage.

        Args:
            display_name : The name of the backend

        Raises:
            FileNotFoundError: If the backend does not exist

        Returns:
            The configuration of the backend in complete form.
        """
        # get the collection on which we work
        _, config_collection = self._get_database_and_collection(self.configs_path)

        # create the filter for the document with display_name that is equal to display_name
        document_to_find = {"display_name": display_name}
        backend_config_dict = config_collection.find_one(document_to_find)

        signed_document_to_find = {"payload.display_name": display_name}
        signed_backend_config_dict = config_collection.find_one(signed_document_to_find)

        # work with the unsigned backend
        if backend_config_dict:
            backend_config_dict.pop("_id")
            return BackendConfigSchemaIn(**backend_config_dict)

        # work with the signed backend this is working normally due to the mongodb API, but to make
        # mypy happy, we have to check if the signed_backend_config_dict is not None
        elif signed_backend_config_dict:
            payload = signed_backend_config_dict["payload"]
            return BackendConfigSchemaIn(**payload)

        raise FileNotFoundError("The backend does not exist for the given storage.")

    def _delete_config(self, display_name: DisplayNameStr) -> bool:
        """
        Delete a config from the storage. This is only intended for test purposes.

        Args:
            display_name: The name of the backend to which we want to upload the job

        Raises:
            FileNotFoundError: If the config does not exist.

        Returns:
            Success if the file was deleted successfully
        """

        config_dict = self.get_config(display_name)
        _, collection = self._get_database_and_collection(self.configs_path)

        if not config_dict.sign:
            document_to_find = {"display_name": display_name}
        else:
            document_to_find = {"payload.display_name": display_name}

        result_found = collection.find_one(document_to_find)
        if result_found is None:
            raise FileNotFoundError(f"the config for {display_name} does not exist.")
        self.delete(self.configs_path, str(result_found["_id"]))

        return True

    def create_job_id(self, display_name: DisplayNameStr, username: str) -> str:
        """
        Create a job id for the job.

        Returns:
            The job id
        """
        return (uuid.uuid4().hex)[:24]

    def _delete_status(
        self, display_name: DisplayNameStr, username: str, job_id: str
    ) -> bool:
        """
        Delete a status from the storage. This is only intended for test purposes.

        Args:
            display_name: The name of the backend to which we want to upload the job
            username: The username of the user that is uploading the job
            job_id: The job_id of the job that we want to upload the status for

        Raises:
            FileNotFoundError: If the status does not exist.

        Returns:
            Success if the file was deleted successfully
        """
        status_json_dir = self.get_attribute_path("status", display_name)

        self.delete(storage_path=status_json_dir, job_id=job_id)
        return True

    def _delete_result(self, display_name: DisplayNameStr, job_id: str) -> bool:
        """
        Delete a result from the storage. This is only intended for test purposes.

        Args:
            display_name: The name of the backend to which we want to upload the job
            username: The username of the user that is uploading the job
            job_id: The job_id of the job that we want to upload the status for

        Raises:
            FileNotFoundError: If the result does not exist.

        Returns:
            Success if the file was deleted successfully
        """
        result_json_dir = self.get_attribute_path("results", display_name, job_id)

        self.delete(storage_path=result_json_dir, job_id=job_id)
        return True

    def upload_public_key(
        self, public_jwk: JWK, display_name: DisplayNameStr, role: PksStr = "backend"
    ) -> None:
        """
        The function that uploads the spooler public JWK to the storage.

        Args:
            public_jwk: The JWK that contains the public key
            display_name : The name of the backend
            role: The role of the public key

        Returns:
            None
        """
        # first make sure that the public key is intended for verification
        if not public_jwk.key_ops == "verify":
            raise ValueError("The key is not intended for verification")

        # make sure that the key does not contain a private key
        if public_jwk.d is not None:
            raise ValueError("The key contains a private key")

        if role == "backend":
            # make sure that the key has the correct kid
            config_dict = self.get_config(display_name)
            if public_jwk.kid != config_dict.kid:
                raise ValueError("The key does not have the correct kid.")

        pks_path = self.get_attribute_path("pks")
        _, collection = self._get_database_and_collection(pks_path)

        # first we have to check if the device already exists in the database
        document_to_find = {"kid": public_jwk.kid}

        result_found = collection.find_one(document_to_find)
        if result_found:
            # update the file
            self.update(
                content_dict=public_jwk.model_dump(),
                storage_path=pks_path,
                job_id=result_found["_id"],
            )
            return

        # in the case of the user this uuid should most likely become identical
        if role == "user":
            self.upload(public_jwk.model_dump(), pks_path, display_name)
        else:
            # if the public key does not exist, we have to create it
            config_id = uuid.uuid4().hex[:24]
            self.upload(public_jwk.model_dump(), pks_path, config_id)

    def get_public_key_from_kid(self, kid: str) -> JWK:
        """
        The function that gets public JWK based on the key id.

        Args:
            kid : The key id of the backend

        Returns:
            JWk : The public JWK object
        """
        # get the database on which we work
        pks_path = self.get_attribute_path("pks")
        _, collection = self._get_database_and_collection(pks_path)

        # create the filter for the document with display_name that is equal to display_name
        document_to_find = {"kid": kid}
        public_jwk_dict = collection.find_one(document_to_find)

        if not public_jwk_dict:
            raise FileNotFoundError("The backend does not exist for the given storage.")

        public_jwk_dict.pop("_id")
        return JWK(**public_jwk_dict)

    def _delete_public_key(self, kid: str) -> bool:
        """
        Delete a public key from the storage. This is only intended for test purposes.

        Args:
            kid: The key id of the public key

        Raises:
            FileNotFoundError: If the public key does not exist.

        Returns:
            Success if the file was deleted successfully
        """
        document_to_find = {"kid": kid}
        # get the database on which we work
        pks_path = self.get_attribute_path("pks")
        _, collection = self._get_database_and_collection(pks_path)

        result_found = collection.find_one(document_to_find)
        if result_found is None:
            raise FileNotFoundError(f"The public key with kid {kid} does not exist")
        self.delete(pks_path, str(result_found["_id"]))
        return True

    def update_in_database(
        self,
        result_dict: ResultDict | None,
        status_msg_dict: StatusMsgDict,
        job_id: str,
        display_name: DisplayNameStr,
        private_jwk: Optional[JWK] = None,
    ) -> None:
        """
        Upload the status and result to the `StorageProvider`.

        The function checks if the reported status of the job has changed to DONE. If so, it will create
        a result json file and move the job json file to the finished folder. It will also update the
        status json file.

        Args:
            result_dict: the dictionary containing the result of the job
            status_msg_dict: the dictionary containing the status message of the job
            job_id: the name of the job
            display_name: the name of the backend
            private_jwk: the private JWK to sign the result with

        Returns:
            None

        Raises:

        """

        job_json_start_dir = self.get_attribute_path("running")
        # check if the job is done or had an error
        if status_msg_dict.status == "DONE":
            # test if the result dict is None
            if result_dict is None:
                raise ValueError(
                    "The 'result_dict' argument cannot be None if the job is done."
                )
            result_uploaded = self.upload_result(
                result_dict, display_name, job_id, private_jwk
            )
            if not result_uploaded:
                raise ValueError("The result was not uploaded successfully.")

            # now move the job out of the running jobs into the finished jobs
            job_finished_json_dir = self.get_attribute_path(
                "finished", display_name=display_name
            )
            self.move(job_json_start_dir, job_finished_json_dir, job_id)

        elif status_msg_dict.status == "ERROR":
            # because there was an error, we move the job to the deleted jobs
            deleted_json_dir = self.get_attribute_path("deleted")
            self.move(job_json_start_dir, deleted_json_dir, job_id)

        # TODO: most likely we should raise an error if the status of the job is not DONE or ERROR

        # and create the status json file
        status_json_dir = self.get_attribute_path("status", display_name)

        try:
            self.update(status_msg_dict.model_dump(), status_json_dir, job_id)
        except FileNotFoundError:
            logging.warning(
                "The status file was missing for %s with job_id %s was missing.",
                display_name,
                job_id,
            )
            self.upload_status(display_name, "", job_id)
            self.update(status_msg_dict.model_dump(), status_json_dir, job_id)

    def get_file_queue(self, storage_path: str) -> list[str]:
        """
        Get a list of documents in the collection of all the queued jobs.

        Args:
            storage_path: Where are we looking for the files.

        Returns:
            A list of files that was found.
        """

        _, collection = self._get_database_and_collection(storage_path)

        # now get the id of all the documents in the collection
        results = collection.find({}, {"_id": 1})
        file_list = []
        for result in results:
            file_list.append(str(result["_id"]))
        return file_list

create_job_id(display_name, username)

Create a job id for the job.

Returns:

Type Description
str

The job id

Source code in src/sqooler/storage_providers/mongodb.py
514
515
516
517
518
519
520
521
def create_job_id(self, display_name: DisplayNameStr, username: str) -> str:
    """
    Create a job id for the job.

    Returns:
        The job id
    """
    return (uuid.uuid4().hex)[:24]

get_attribute_id(attribute_name, job_id, display_name=None)

Get the path to the id of the device.

Parameters:

Name Type Description Default
attribute_name AttributeIdStr

The name of the attribute

required
job_id str

The job_id of the job

required
display_name Optional[DisplayNameStr]

The name of the backend

None

Returns:

Type Description
str

The path to the results of the device.

Source code in src/sqooler/storage_providers/mongodb.py
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
def get_attribute_id(
    self,
    attribute_name: AttributeIdStr,
    job_id: str,
    display_name: Optional[DisplayNameStr] = None,
) -> str:
    """
    Get the path to the id of the device.

    Args:
        attribute_name: The name of the attribute
        job_id: The job_id of the job
        display_name: The name of the backend

    Returns:
        The path to the results of the device.
    """

    match attribute_name:
        case "configs":
            raise ValueError(f"The attribute name {attribute_name} is not valid.")
        case "job":
            _id = job_id
        case "results":
            _id = job_id
        case "status":
            _id = job_id
        case _:
            raise ValueError(f"The attribute name {attribute_name} is not valid.")
    return _id

get_attribute_path(attribute_name, display_name=None, job_id=None, username=None)

Get the path to the results of the device.

Parameters:

Name Type Description Default
display_name Optional[DisplayNameStr]

The name of the backend

None
attribute_name AttributePathStr

The name of the attribute

required
job_id Optional[str]

The job_id of the job

None
username Optional[str]

The username of the user

None

Returns:

Type Description
str

The path to the results of the device.

Source code in src/sqooler/storage_providers/mongodb.py
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
def get_attribute_path(
    self,
    attribute_name: AttributePathStr,
    display_name: Optional[DisplayNameStr] = None,
    job_id: Optional[str] = None,
    username: Optional[str] = None,
) -> str:
    """
    Get the path to the results of the device.

    Args:
        display_name: The name of the backend
        attribute_name: The name of the attribute
        job_id: The job_id of the job
        username: The username of the user

    Returns:
        The path to the results of the device.
    """

    match attribute_name:
        case "configs":
            path = self.configs_path
        case "results":
            path = f"{self.results_path}/{display_name}"
        case "running":
            path = self.running_path
        case "status":
            path = f"{self.status_path}/{display_name}"
        case "queue":
            path = f"{self.queue_path}/{display_name}"
        case "deleted":
            path = self.deleted_path
        case "finished":
            path = f"{self.finished_path}/{display_name}"
        case "pks":
            path = self.pks_path
        case _:
            raise ValueError(f"The attribute name {attribute_name} is not valid.")
    return path

get_backends()

Get a list of all the backends that the provider offers.

Source code in src/sqooler/storage_providers/mongodb.py
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
@validate_active
def get_backends(self) -> list[DisplayNameStr]:
    """
    Get a list of all the backends that the provider offers.
    """

    # get the collection on which we work
    _, config_collection = self._get_database_and_collection(self.configs_path)

    # get all the documents in the collection configs and save the disply_name in a list
    backend_names: list[DisplayNameStr] = []
    for config_dict in config_collection.find():
        config_dict.pop("_id")
        expected_keys_for_jws = {"header", "payload", "signature"}
        if set(config_dict.keys()) == expected_keys_for_jws:
            backend_names.append(config_dict["payload"]["display_name"])
        else:
            backend_names.append(config_dict["display_name"])
    return backend_names

get_config(display_name)

The function that downloads the spooler configuration to the storage.

Parameters:

Name Type Description Default
display_name

The name of the backend

required

Raises:

Type Description
FileNotFoundError

If the backend does not exist

Returns:

Type Description
BackendConfigSchemaIn

The configuration of the backend in complete form.

Source code in src/sqooler/storage_providers/mongodb.py
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
@validate_active
def get_config(self, display_name: DisplayNameStr) -> BackendConfigSchemaIn:
    """
    The function that downloads the spooler configuration to the storage.

    Args:
        display_name : The name of the backend

    Raises:
        FileNotFoundError: If the backend does not exist

    Returns:
        The configuration of the backend in complete form.
    """
    # get the collection on which we work
    _, config_collection = self._get_database_and_collection(self.configs_path)

    # create the filter for the document with display_name that is equal to display_name
    document_to_find = {"display_name": display_name}
    backend_config_dict = config_collection.find_one(document_to_find)

    signed_document_to_find = {"payload.display_name": display_name}
    signed_backend_config_dict = config_collection.find_one(signed_document_to_find)

    # work with the unsigned backend
    if backend_config_dict:
        backend_config_dict.pop("_id")
        return BackendConfigSchemaIn(**backend_config_dict)

    # work with the signed backend this is working normally due to the mongodb API, but to make
    # mypy happy, we have to check if the signed_backend_config_dict is not None
    elif signed_backend_config_dict:
        payload = signed_backend_config_dict["payload"]
        return BackendConfigSchemaIn(**payload)

    raise FileNotFoundError("The backend does not exist for the given storage.")

get_file_queue(storage_path)

Get a list of documents in the collection of all the queued jobs.

Parameters:

Name Type Description Default
storage_path str

Where are we looking for the files.

required

Returns:

Type Description
list[str]

A list of files that was found.

Source code in src/sqooler/storage_providers/mongodb.py
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
def get_file_queue(self, storage_path: str) -> list[str]:
    """
    Get a list of documents in the collection of all the queued jobs.

    Args:
        storage_path: Where are we looking for the files.

    Returns:
        A list of files that was found.
    """

    _, collection = self._get_database_and_collection(storage_path)

    # now get the id of all the documents in the collection
    results = collection.find({}, {"_id": 1})
    file_list = []
    for result in results:
        file_list.append(str(result["_id"]))
    return file_list

get_public_key_from_kid(kid)

The function that gets public JWK based on the key id.

Parameters:

Name Type Description Default
kid

The key id of the backend

required

Returns:

Name Type Description
JWk JWK

The public JWK object

Source code in src/sqooler/storage_providers/mongodb.py
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
def get_public_key_from_kid(self, kid: str) -> JWK:
    """
    The function that gets public JWK based on the key id.

    Args:
        kid : The key id of the backend

    Returns:
        JWk : The public JWK object
    """
    # get the database on which we work
    pks_path = self.get_attribute_path("pks")
    _, collection = self._get_database_and_collection(pks_path)

    # create the filter for the document with display_name that is equal to display_name
    document_to_find = {"kid": kid}
    public_jwk_dict = collection.find_one(document_to_find)

    if not public_jwk_dict:
        raise FileNotFoundError("The backend does not exist for the given storage.")

    public_jwk_dict.pop("_id")
    return JWK(**public_jwk_dict)

update_config(config_dict, display_name, private_jwk=None)

The function that updates the spooler configuration on the storage.

Parameters:

Name Type Description Default
config_dict BackendConfigSchemaIn

The dictionary containing the configuration

required
display_name

The name of the backend

required
private_jwk Optional[JWK]

The private key of the backend

None

Returns:

Type Description
None

None

Source code in src/sqooler/storage_providers/mongodb.py
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
def update_config(
    self,
    config_dict: BackendConfigSchemaIn,
    display_name: DisplayNameStr,
    private_jwk: Optional[JWK] = None,
) -> None:
    """
    The function that updates the spooler configuration on the storage.

    Args:
        config_dict: The dictionary containing the configuration
        display_name : The name of the backend
        private_jwk: The private key of the backend

    Returns:
        None
    """

    config_dict = self._verify_config(config_dict, display_name)

    _, collection = self._get_database_and_collection(self.configs_path)

    # now make sure that we add the timezone as we open the file
    collection_with_tz = collection.with_options(
        codec_options=CodecOptions(tz_aware=True, tzinfo=timezone.utc)
    )
    # first we have to check if the device already exists in the database
    document_to_find = {"display_name": display_name}
    result_found = collection_with_tz.find_one(document_to_find)

    signed_document_to_find = {"payload.display_name": display_name}
    signed_backend_config_dict = collection_with_tz.find_one(
        signed_document_to_find
    )

    if result_found:
        old_config_jws = result_found
        job_id = result_found["_id"]
    elif signed_backend_config_dict:
        old_config_jws = signed_backend_config_dict
        job_id = signed_backend_config_dict["_id"]
        old_config_jws.pop("_id")
    else:
        raise FileNotFoundError(
            (
                f"The config for {display_name} does not exist and should not be updated."
                "Use the upload_config method instead."
            )
        )
    upload_dict = self._format_update_config(
        old_config_jws, config_dict, private_jwk
    )

    self.update(
        content_dict=upload_dict,
        storage_path=self.configs_path,
        job_id=job_id,
    )

update_in_database(result_dict, status_msg_dict, job_id, display_name, private_jwk=None)

Upload the status and result to the StorageProvider.

The function checks if the reported status of the job has changed to DONE. If so, it will create a result json file and move the job json file to the finished folder. It will also update the status json file.

Parameters:

Name Type Description Default
result_dict ResultDict | None

the dictionary containing the result of the job

required
status_msg_dict StatusMsgDict

the dictionary containing the status message of the job

required
job_id str

the name of the job

required
display_name DisplayNameStr

the name of the backend

required
private_jwk Optional[JWK]

the private JWK to sign the result with

None

Returns:

Type Description
None

None

Raises:

Source code in src/sqooler/storage_providers/mongodb.py
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
def update_in_database(
    self,
    result_dict: ResultDict | None,
    status_msg_dict: StatusMsgDict,
    job_id: str,
    display_name: DisplayNameStr,
    private_jwk: Optional[JWK] = None,
) -> None:
    """
    Upload the status and result to the `StorageProvider`.

    The function checks if the reported status of the job has changed to DONE. If so, it will create
    a result json file and move the job json file to the finished folder. It will also update the
    status json file.

    Args:
        result_dict: the dictionary containing the result of the job
        status_msg_dict: the dictionary containing the status message of the job
        job_id: the name of the job
        display_name: the name of the backend
        private_jwk: the private JWK to sign the result with

    Returns:
        None

    Raises:

    """

    job_json_start_dir = self.get_attribute_path("running")
    # check if the job is done or had an error
    if status_msg_dict.status == "DONE":
        # test if the result dict is None
        if result_dict is None:
            raise ValueError(
                "The 'result_dict' argument cannot be None if the job is done."
            )
        result_uploaded = self.upload_result(
            result_dict, display_name, job_id, private_jwk
        )
        if not result_uploaded:
            raise ValueError("The result was not uploaded successfully.")

        # now move the job out of the running jobs into the finished jobs
        job_finished_json_dir = self.get_attribute_path(
            "finished", display_name=display_name
        )
        self.move(job_json_start_dir, job_finished_json_dir, job_id)

    elif status_msg_dict.status == "ERROR":
        # because there was an error, we move the job to the deleted jobs
        deleted_json_dir = self.get_attribute_path("deleted")
        self.move(job_json_start_dir, deleted_json_dir, job_id)

    # TODO: most likely we should raise an error if the status of the job is not DONE or ERROR

    # and create the status json file
    status_json_dir = self.get_attribute_path("status", display_name)

    try:
        self.update(status_msg_dict.model_dump(), status_json_dir, job_id)
    except FileNotFoundError:
        logging.warning(
            "The status file was missing for %s with job_id %s was missing.",
            display_name,
            job_id,
        )
        self.upload_status(display_name, "", job_id)
        self.update(status_msg_dict.model_dump(), status_json_dir, job_id)

upload_config(config_dict, display_name, private_jwk=None)

The function that uploads the spooler configuration to the storage.

Parameters:

Name Type Description Default
config_dict BackendConfigSchemaIn

The dictionary containing the configuration

required
display_name

The name of the backend

required
private_jwk Optional[JWK]

The private JWK to sign the configuration with

None

Returns:

Type Description
None

None

Source code in src/sqooler/storage_providers/mongodb.py
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
def upload_config(
    self,
    config_dict: BackendConfigSchemaIn,
    display_name: DisplayNameStr,
    private_jwk: Optional[JWK] = None,
) -> None:
    """
    The function that uploads the spooler configuration to the storage.

    Args:
        config_dict: The dictionary containing the configuration
        display_name : The name of the backend
        private_jwk: The private JWK to sign the configuration with

    Returns:
        None
    """
    config_dict = self._verify_config(config_dict, display_name)

    # first we have to check if the device already exists in the database

    document_to_find = {"display_name": display_name}

    # get the collection on which we work
    _, collection = self._get_database_and_collection(self.configs_path)

    document_to_find = {"display_name": display_name}
    result_found = collection.find_one(document_to_find)
    if result_found:
        raise FileExistsError(
            f"The configuration for {display_name} already exists and should not be overwritten."
        )

    # now also look for signed configurations
    signed_document_to_find = {"payload.display_name": display_name}
    result_found = collection.find_one(signed_document_to_find)
    if result_found:
        raise FileExistsError(
            f"The configuration for {display_name} already exists and should not be overwritten."
        )

    upload_dict = self._format_config_dict(config_dict, private_jwk)
    config_id = uuid.uuid4().hex[:24]
    self.upload(upload_dict, self.configs_path, config_id)

upload_public_key(public_jwk, display_name, role='backend')

The function that uploads the spooler public JWK to the storage.

Parameters:

Name Type Description Default
public_jwk JWK

The JWK that contains the public key

required
display_name

The name of the backend

required
role PksStr

The role of the public key

'backend'

Returns:

Type Description
None

None

Source code in src/sqooler/storage_providers/mongodb.py
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
def upload_public_key(
    self, public_jwk: JWK, display_name: DisplayNameStr, role: PksStr = "backend"
) -> None:
    """
    The function that uploads the spooler public JWK to the storage.

    Args:
        public_jwk: The JWK that contains the public key
        display_name : The name of the backend
        role: The role of the public key

    Returns:
        None
    """
    # first make sure that the public key is intended for verification
    if not public_jwk.key_ops == "verify":
        raise ValueError("The key is not intended for verification")

    # make sure that the key does not contain a private key
    if public_jwk.d is not None:
        raise ValueError("The key contains a private key")

    if role == "backend":
        # make sure that the key has the correct kid
        config_dict = self.get_config(display_name)
        if public_jwk.kid != config_dict.kid:
            raise ValueError("The key does not have the correct kid.")

    pks_path = self.get_attribute_path("pks")
    _, collection = self._get_database_and_collection(pks_path)

    # first we have to check if the device already exists in the database
    document_to_find = {"kid": public_jwk.kid}

    result_found = collection.find_one(document_to_find)
    if result_found:
        # update the file
        self.update(
            content_dict=public_jwk.model_dump(),
            storage_path=pks_path,
            job_id=result_found["_id"],
        )
        return

    # in the case of the user this uuid should most likely become identical
    if role == "user":
        self.upload(public_jwk.model_dump(), pks_path, display_name)
    else:
        # if the public key does not exist, we have to create it
        config_id = uuid.uuid4().hex[:24]
        self.upload(public_jwk.model_dump(), pks_path, config_id)

Comments