Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the input file #2

Open
Lily-WL opened this issue Jun 21, 2016 · 1 comment
Open

About the input file #2

Lily-WL opened this issue Jun 21, 2016 · 1 comment

Comments

@Lily-WL
Copy link

Lily-WL commented Jun 21, 2016

Dear researcher,

In the paper of Dick (2009), it mentioned that the input fasta containing contigs > 2kb. And the contigs longer than 5 kb was split into shorter. I just want to inquire if you can suggest some software to prepare such kind of input data.

Best!

@MehdinMasinovic
Copy link

Hello @Lily-WL ,
I am currently also using tetraESOM. I believe that you don't need to prepare the input data yourself, the package does it on its own.

If you place your contigs into one folder (as described by the manual) and run the esomWrapper.pl script, it will create one file called Tetra_yourprefix_2500_5000_split.fasta

The "yourprefix" is the prefix you provide when running the esomWrapper.pl script.

In my dataset, for example, my first contig is ~100kb long and is called "contig_100_0". If I now check the Tetra_..._split.fasta file, I will see this contig split into several parts, like here:

contig_100_0_1
Sequence
contig_100_0_2
Sequence
...

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants