Hi Everyone, as i made a few 16S workflow i was asked to test Galaxy and see its strengths and weaknesses. I run my pipeline on my 36 samples and it failed due to to much information, which I can understand I have more than 12 millions sequences on my samples. So I reduced my number of samples to 13 which make 2 millions sequences, but at a specific step ( pre.cluster ) , it doesn't tell me an error, but keep going and for now more than 15h !! I'm wondering is it still to much ? How can i Know that something is going on and that i m not wasting my time here ! Thank you very much for your answer !
I assume you are working on usegalaxy.org.
I do not know what tool you are executing that takes this much time, but it is possible for some tools to take this long with large inputs. Generally if the tool runs into trouble it will stop and show an error that you would see in Galaxy interface.