[Bioperl-pipeline] Still working with biopipe....
matthieu CONTE
m_conte at hotmail.com
Wed Nov 26 08:26:10 EST 2003
I still have problem with my pipeline blast's program.....
"
Creating biopipe
Loading Schema...
Reading Data_setup xml : /home/conte/xml/newhope.xml
Doing DBAdaptor and IOHandler setup
Doing Transformers..
Doing Pipeline Flow Setup
Doing Analysis..
Doing Rules
Doing Job Setup...
Loading of pipeline biopipe completed
2 analysis found.
Running test and setup..
//////////// Analysis Test ////////////
Checking Analysis 1 DataMonger
-------------------- WARNING ---------------------
MSG: Skipping test for DataMonger
---------------------------------------------------
ok
Checking Analysis 2 Blast ok
Fetching Jobs...
Fetched 1 incomplete jobs
Running job /tmp//6/biopipe_DataMonger.1069852072.635.out
/tmp//6/biopipe_DataMonger.1069852072.635.err
"
I think there is a problem with the method "get_Seq_by_id"
I make a little bioperl program to test the use of the association of the
"get_all_ids" and "get_Seq_by_ids" it just return the ids and not sequences
only the method "seq" give the sequences but I didn't manage to use it in my
XML code!.....
use Bio::DB::Fasta;
my $index_file_name='oriz_mfasta.txt';
my $inx=Bio::DB::Fasta->new($index_file_name);
my @ids=$inx->get_all_ids();
foreach (@ids)
{
my $seq=$inx->seq($_);
print "$_!!!!$seq\n";
}
Matthieu CONTE
M. Sc. in Bioinformatics form SIB
00 33 06.68.90.28.70
m_conte at hotmail.com
>From: Shawn Hoon <shawnh at stanford.edu>
>To: "matthieu CONTE" <m_conte at hotmail.com>
>CC: bioperl-pipeline at bioperl.org
>Subject: Re: [Bioperl-pipeline] Still working with biopipe....
>Date: Fri, 21 Nov 2003 10:01:22 -0800
>
>Hi Mattieu,
> Thanks for trying this out.
>are you using the bioperl-pipeline bundle from the website? It looks like
>it and that is the correct version to use.
>There are some problems with you xml file mainly you need to put in the
>attribute ids for iohandler/analysis etc.
>I have attached a corrected version of your xml file. You should try and
>look at bioperl-pipeline/xml/examples/xml/blast_db_flat.xml
>in the examples directory and use that as a template.
>
>
>hope that helps
>
>shawn
><< test.xml >>
>
>
>On Friday, November 21, 2003, at 7:44AM, matthieu CONTE wrote:
>
>>Hi,
>>Still working with biopipe....
>>Im now trying to create de novo a pipeline to find orthologues between
>>Oryza sativa (Os) and Arabidopsis thaliana (At) by BBMH (best blast mutal
>>hit) (before to develop something more efficient and more complicate !).
>>So I started by a simple blast between a prot from Os to At multifasta
>>prot using the Bio::DB::Fasta and all the bioperl methods needed (and
>>loop on all the Os proteins instead of a massive blast with a chunk of Os
>>proteins). I would like to take a sequence from oriz_mfasta.txt ( using
>>the get_Seq_by_id fonction) and blast it against arabido_mfasta.txt and
>>so on for all the seq of oryza.This is the first step. But, it's not
>>working !!! Probably because it's not really clear for me the function of
>>all the XML code I am working with (especially the <datamonger> tag !).
>>
>>You will find the code and the biopipe output below.
>>Thanks in advance
>>
>>
>>
>><pipeline_setup>
>>
>><!-- FILES -->
>><global
>> rootdir="/home/conte/test_blast"
>> datadir="$rootdir/datahope"
>> workdir="$rootdir/blasthope"
>> inputfile="$datadir/oriz_mfasta.txt"
>> blastpath = ""
>> blast_param1="-p blastp -e 1e-5"
>> blastdb1="$datadir/arabido_mfasta.txt"
>> resultdir1="$rootdir/resulthope/analysis1"
>>/>
>><pipeline_flow_setup>
>><!--CALL MODULES -->
>> <database_setup>
>> <streamadaptor>
>> <module>Bio::Pipeline::Dumper</module>
>> </streamadaptor>
>> <streamadaptor>
>> <module>Bio::DB::Fasta</module>
>> </streamadaptor>
>> </database_setup>
>>
>><!-- IOHANDLER PICK UP iDs-->
>> <iohandler_setup>
>> <iohandler>
>> <adaptor_id>2</adaptor_id>
>> <adaptor_type>STREAM</adaptor_type>
>> <iohandler_type>INPUT</iohandler_type>
>> <method>
>> <name>new</name>
>> <rank>1</rank>
>> <argument>
>> <value>$inputfile</value>
>> </argument>
>> </method>
>> <method>
>> <name>get_Seq_by_id</name>
>> <argument>
>> <value>INPUT</value>
>> </argument>
>> <rank>2</rank>
>> </method>
>> </iohandler>
>>
>> <iohandler>
>> <adaptor_id>2</adaptor_id>
>> <adaptor_type>STREAM</adaptor_type>
>> <iohandler_type>INPUT</iohandler_type>
>> <method>
>> <name>new</name>
>> <rank>1</rank>
>> <argument>
>> <value>$inputfile</value>
>> </argument>
>> </method>
>> <method>
>> <name>get_all_ids</name>
>> <rank>2</rank>
>> </method>
>> </iohandler>
>>
>><!-- PARAMETRES OUTPUT (DUMPER) -->
>> <iohandler>
>> <adaptor_id>1</adaptor_id>
>> <adaptor_type>STREAM</adaptor_type>
>> <iohandler_type>OUTPUT</iohandler_type>
>> <method>
>> <name>new</name>
>> <rank>1</rank>
>> <argument>
>> <tag>-dir</tag>
>> <value>$resultdir1</value>
>> SCALAR
>> <rank>1</rank>
>> </argument>
>> <argument>
>> <tag>-module</tag>
>> <value>generic</value>
>> SCALAR
>> <rank>1</rank>
>> </argument>
>> <argument>
>> <tag>-prefix</tag>
>> SCALAR
>> <value>INPUT</value>
>> <rank>2</rank>
>> </argument>
>> <argument>
>> <tag>-format</tag>
>> SCALAR
>> <value>gff</value>
>> <rank>3</rank>
>> </argument>
>> <argument>
>> <tag>-file_suffix</tag>
>> SCALAR
>> <value>gff</value>
>> <rank>4</rank>
>> </argument>
>> </method>
>> <method>
>> <name>dump</name>
>> <rank>2</rank>
>> <argument>
>> <value>OUTPUT</value>
>> ARRAY
>> <rank>1</rank>
>> </argument>
>> </method>
>> </iohandler>
>> </iohandler_setup>
>>
>><!-- ANALYSIS -->
>> <analysis>
>> <data_monger>
>> <initial></initial>
>> <input>
>> <name>protein_ids</name>
>> <iohandler>1</iohandler>
>> </input>
>> <input_create>
>> <module>setup_initial</module>
>> <rank>1</rank>
>> <argument>
>> <tag>protein_ids</tag>
>> <value>2</value>
>> </argument>
>> </input_create>
>></data_monger>
>><input_iohandler></input_iohandler>
>> </analysis>
>>
>><!-- BLAST-->
>> <analysis>
>> <logic_name>Blast</logic_name>
>> <runnable>Bio::Pipeline::Runnable::Blast</runnable>
>> <db>family</db>
>> <db_file>$blastdb1</db_file>
>> <program>blastall</program>
>>
>><!-- BLASTPATH-->
>> <program_file>$blastpath</program_file>
>> <analysis_parameters>$blast_param1</analysis_parameters>
>> <runnable_parameters>-formatdb 1 -result_dir
>>$resultdir1</runnable_parameters>
>>
>> <input_iohandler></input_iohandler>
>>
>> <output_iohandler></output_iohandler>
>> </analysis>
>>
>><!-- RULES -->
>><rule>
>> <current_analysis_id>1</current_analysis_id>
>> <next_analysis_id>2</next_analysis_id>
>> NOTHING
>>
>></rule>
>>
>></pipeline_flow_setup>
>><job_setup>
>></job_setup>
>>
>></pipeline_setup>
>>
>>
>>And I obtain:
>>
>>Creating biopipe
>> Loading Schema...
>>Reading Data_setup xml : /home/conte/xml/newhope.xml
>>Doing DBAdaptor and IOHandler setup
>>Doing Pipeline Flow Setup
>>Doing Analysis..
>>
>>------------- EXCEPTION -------------
>>MSG: Need to store analysis first
>>STACK Bio::Pipeline::SQL::JobAdaptor::store
>>/usr/local/ActivePerl-5.8/lib/site_perl/5.8.0/Bio/Pipeline/SQL/
>>JobAdaptor.pm:459
>>STACK Bio::Pipeline::XMLImporter::_create_initial_input_and_job
>>/usr/local/ActivePerl-5.8/lib/site_perl/5.8.0/Bio/Pipeline/
>>XMLImporter.pm:837
>>STACK Bio::Pipeline::XMLImporter::run
>>/usr/local/ActivePerl-5.8/lib/site_perl/5.8.0/Bio/Pipeline/
>>XMLImporter.pm:484
>>STACK toplevel PipelineManager:120
>>
>>
>>
>>
>>
>>
>>
>>
>>Matthieu CONTE
>>M. Sc. in Bioinformatics form SIB
>>CIRAD
>>00 33 06.68.90.28.70
>>m_conte at hotmail.com
>>
>>_________________________________________________________________
>>MSN Search, le moteur de recherche qui pense comme vous !
>>http://search.msn.fr/worldwide.asp
>>
>>_______________________________________________
>>bioperl-pipeline mailing list
>>bioperl-pipeline at bioperl.org
>>http://bioperl.org/mailman/listinfo/bioperl-pipeline
_________________________________________________________________
MSN Messenger : discutez en direct avec vos amis !
http://www.msn.fr/msger/default.asp
More information about the bioperl-pipeline
mailing list