Making large AI models cheaper, faster and more accessible
The Transformer architecture has improved the performance of deep learning models in domains such as Computer Vision and Natural Language Processing. Together with better performance come larger model sizes. This imposes challenges to the memory wall of the current accelerator hardware such as GPU. It is never ideal to train large models such as Vision Transformer, BERT, and GPT on a single GPU or a single machine. There is an urgent demand to train models in a distributed environment....
CometAnalyser, for quantitative comet assay analysis.
Description:
Comet assay provides an easy solution to estimate DNA damage in single cells through microscopy assessment. To obtain reproducible and reliable quantitative data, we developed an easy-to-use tool named CometAnalyser. CometAnalyser is an open-source deep-learning tool designed for the analysis of both fluorescent and silver-stained wide-field microscopy images. Once the comets are segmented and classified, several intensity/morphological features are automatically exported as a spreadsheet file.
...