Details
-
Feature
-
Should have
-
SRCnet
-
-
-
1
-
1
-
0
-
Team_MAGENTA
-
Sprint 5
-
-
-
-
21.6
-
Stories Completed, Demonstrated, Satisfies Acceptance Criteria, Accepted by FO
-
-
Team_Magenta
Description
All of the current tasks in the workloads Gitlab repository use small/minimal datasets to demonstrate the task. Moving forward we want to scale some of these up to use much larger datasets. This can be to test the task can handle significant datasets, but also to test the compute infrastructure running the task. This will significantly increase the compute time, and will help test more realistic workloads being run at SRCs. Note that this is not just running the same task over and over again, but using a significantly larger dataset.
This feature would selection at least one of the tasks, and scale it up. It could be split into multiple features, per tasks that are relevant to scale up.
Run PYBDSF source-finding on a significantly larger set of images.
Mosaicking large areas of sky.
Image-cutouts for thousands/millions of sources from multiple images.
Image convolution - convolving many images to alter the resolution.
CNN image classifier - use a much larger dataset of images.
Develop a script that runs all tasks in their default state, to test the functionality of an SRC node.