<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://hivelab.biochemistry.gwu.edu/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Christie.woodside</id>
	<title>HIVE Lab - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://hivelab.biochemistry.gwu.edu/wiki/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Christie.woodside"/>
	<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/Special:Contributions/Christie.woodside"/>
	<updated>2026-04-30T21:14:09Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.42.1</generator>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=1234</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=1234"/>
		<updated>2026-04-07T19:36:09Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{DISPLAYTITLE:&amp;lt;span style=&amp;quot;position:absolute; clip:rect(1px, 1px, 1px, 1px);&amp;quot;&amp;gt;{{FULLPAGENAME}}&amp;lt;/span&amp;gt;}}&lt;br /&gt;
__NOTOC__&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- TOP BANNER --&amp;gt;&lt;br /&gt;
&amp;lt;div id=&amp;quot;argos-topbanner&amp;quot; style=&amp;quot;clear:both; position:relative; box-sizing:border-box; width:100%; margin:1.2em 0 8px; border:1px solid #ddd; background-color:#f9f9f9; color:#000;&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;div style=&amp;quot;margin:0.6em; text-align:center;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;font-size:170%; padding:.1em;&amp;quot;&amp;gt;Welcome to the FDA-ARGOS Wiki&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;font-size:100%;&amp;quot;&amp;gt;&lt;br /&gt;
      This wiki provides project information, database resources, protocols, publications, and support materials for the FDA-ARGOS initiative.&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
  &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;div style=&amp;quot;clear:both;&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- INTRODUCTION --&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;margin:10px 0; border:1px solid #CCC; padding:0 12px 12px 12px; box-shadow:0 2px 2px rgba(0,0,0,0.1); background:#f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;h2&amp;gt;Introduction&amp;lt;/h2&amp;gt;&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see:&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- FDA-ARGOS DATABASE --&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;margin:10px 0; border:1px solid #CCC; padding:0 12px 12px 12px; box-shadow:0 2px 2px rgba(0,0,0,0.1); background:#f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;h2&amp;gt;FDA-ARGOS Database&amp;lt;/h2&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
The organisms in the database tables that have FDA-ARGOS tags can be found in our FDA-ARGOS NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 here]. We also have QC&#039;d and included additional organisms that were publicly available on NCBI. They are logged in this page, [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
A comprehensive reference table of human pathogenic organisms, including bacteria, viruses, and eukaryotic pathogens, has been created. The table integrates data from multiple curated sources and includes taxonomic identifiers and classification information for downstream analysis and database integration.&lt;br /&gt;
&lt;br /&gt;
* [[Comprehensive Pathogenic Organisms Reference]]&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here]. This is a static webpage that will be updated with each push.&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- FDA-ARGOS FAQS --&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;margin:10px 0; border:1px solid #CCC; padding:0 12px 12px 12px; box-shadow:0 2px 2px rgba(0,0,0,0.1); background:#f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;h2&amp;gt;FDA-ARGOS FAQs&amp;lt;/h2&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- ARGOS USAGE TUTORIALS AND PROTOCOLS --&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;margin:10px 0; border:1px solid #CCC; padding:0 12px 12px 12px; box-shadow:0 2px 2px rgba(0,0,0,0.1); background:#f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;h2&amp;gt;ARGOS Usage Tutorials and Protocols&amp;lt;/h2&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The ARGOS One-Click Pipeline is used to create the QC metrics and results displayed in our data tables on data.argosdb.org. The ARGOS One-Click Pipeline Usage tutorial can be found [[ARGOSQC Usage Tutorial|here]].&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- ROW 4 --&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;display:flex; flex-flow:row wrap; justify-content:space-between; padding:0; margin:0 -5px 0 -5px;&amp;quot;&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;div style=&amp;quot;flex:1; margin:5px; min-width:300px; border:1px solid #CCC; padding:0 12px 12px 12px; box-shadow:0 2px 2px rgba(0,0,0,0.1); background:#f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;h2&amp;gt;Project Publications&amp;lt;/h2&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nature Communications&#039;&#039; 10, 3313 (2019). [https://doi.org/10.1038/s41467-019-11306-6 DOI: 10.1038/s41467-019-11306-6]&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;Coming soon:&#039;&#039; Journal publication highlighting QC metrics and the data dictionary for the ARGOS database.&lt;br /&gt;
  &amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=1233</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=1233"/>
		<updated>2026-04-07T19:33:58Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{DISPLAYTITLE:&amp;lt;span style=&amp;quot;position:absolute; clip:rect(1px, 1px, 1px, 1px);&amp;quot;&amp;gt;{{FULLPAGENAME}}&amp;lt;/span&amp;gt;}}&lt;br /&gt;
__NOTOC__&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- TOP BANNER --&amp;gt;&lt;br /&gt;
&amp;lt;div id=&amp;quot;argos-topbanner&amp;quot; style=&amp;quot;clear:both; position:relative; box-sizing:border-box; width:100%; margin:1.2em 0 8px; border:1px solid #ddd; background-color:#f9f9f9; color:#000;&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;div style=&amp;quot;margin:0.6em; text-align:center;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;font-size:170%; padding:.1em;&amp;quot;&amp;gt;Welcome to the FDA-ARGOS Wiki&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;font-size:100%;&amp;quot;&amp;gt;&lt;br /&gt;
      This wiki provides project information, database resources, protocols, publications, and support materials for the FDA-ARGOS initiative.&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
  &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The organisms in the database tables that have FDA-ARGOS tags can be found in our FDA-ARGOS NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 here]. We also have QC&#039;d and included additional organisms that were publicly available on NCBI. They are logged in this page, [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
A comprehensive reference table of human pathogenic organisms, including bacteria, viruses, and eukaryotic pathogens, has been created. The table integrates data from multiple curated sources and includes taxonomic identifiers and classification information for downstream analysis and database integration. &lt;br /&gt;
&lt;br /&gt;
* [[Comprehensive Pathogenic Organisms Reference]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here]. This is a static webpage that will be updated with each push.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&lt;br /&gt;
== ARGOS Usage Tutorials and Protocols ==&lt;br /&gt;
&lt;br /&gt;
The ARGOS One-Click Pipeline is used to create the QC metrics and results displayed in our data tables on data.argosdb.org. The ARGOS One-Click Pipeline Usage tutorial can be found [[ARGOSQC Usage Tutorial|here]]. &lt;br /&gt;
Resources:&lt;br /&gt;
* [[ARGOSQC Usage Tutorial|ARGOS One-Click Pipeline Usage Tutorial]]&lt;br /&gt;
  &amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;!-- ROW 4 --&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;display:flex; flex-flow:row wrap; justify-content:space-between; padding:0; margin:0 -5px 0 -5px;&amp;quot;&amp;gt;&lt;br /&gt;
&lt;br /&gt;
  &amp;lt;div style=&amp;quot;flex:1; margin:5px; min-width:300px; border:1px solid #CCC; padding:0 12px 12px 12px; box-shadow:0 2px 2px rgba(0,0,0,0.1); background:#f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;h2&amp;gt;Project Publications&amp;lt;/h2&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nature Communications&#039;&#039; 10, 3313 (2019). [https://doi.org/10.1038/s41467-019-11306-6 DOI: 10.1038/s41467-019-11306-6]&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;Coming soon:&#039;&#039; Journal publication highlighting QC metrics and the data dictionary for the ARGOS database.&lt;br /&gt;
  &amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=1232</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=1232"/>
		<updated>2026-04-07T19:16:47Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* FDA-ARGOS Database */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The organisms in the database tables that have FDA-ARGOS tags can be found in our FDA-ARGOS NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 here]. We also have QC&#039;d and included additional organisms that were publicly available on NCBI. They are logged in this page, [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
A comprehensive reference table of human pathogenic organisms, including bacteria, viruses, and eukaryotic pathogens, has been created. The table integrates data from multiple curated sources and includes taxonomic identifiers and classification information for downstream analysis and database integration. &lt;br /&gt;
&lt;br /&gt;
* [[Comprehensive Pathogenic Organisms Reference]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here]. This is a static webpage that will be updated with each push.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&lt;br /&gt;
== ARGOS Usage Tutorials and Protocols ==&lt;br /&gt;
&lt;br /&gt;
The ARGOS One-Click Pipeline is used to create the QC metrics and results displayed in our data tables on data.argosdb.org. The ARGOS One-Click Pipeline Usage tutorial can be found [[ARGOSQC Usage Tutorial|here]]. &lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&lt;br /&gt;
*Coming Soon: Journal Publication highlighting our QC metrics and data dictionary for the ARGOS Database&amp;lt;br /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=1231</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=1231"/>
		<updated>2026-04-07T19:15:39Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* FDA-ARGOS Database */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The organisms in the database tables that have FDA-ARGOS tags can be found in our FDA-ARGOS NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 here]. We also have QC&#039;d and included additional organisms that were publicly available on NCBI. They are logged in this page, [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
* [[Comprehensive Pathogenic Organisms Reference]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here]. This is a static webpage that will be updated with each push.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&lt;br /&gt;
== ARGOS Usage Tutorials and Protocols ==&lt;br /&gt;
&lt;br /&gt;
The ARGOS One-Click Pipeline is used to create the QC metrics and results displayed in our data tables on data.argosdb.org. The ARGOS One-Click Pipeline Usage tutorial can be found [[ARGOSQC Usage Tutorial|here]]. &lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&lt;br /&gt;
*Coming Soon: Journal Publication highlighting our QC metrics and data dictionary for the ARGOS Database&amp;lt;br /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Spring_2026&amp;diff=1142</id>
		<title>Volunteership Spring 2026</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Spring_2026&amp;diff=1142"/>
		<updated>2026-01-26T15:47:31Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Volunteers (TBD) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== 2026 Spring Volunteer Program Details ==&lt;br /&gt;
&lt;br /&gt;
=== Dates ===&lt;br /&gt;
&#039;&#039;&#039;Application Deadline&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
January 9, 2026, Noon (email your updated resume and projects in order of preference). Acceptance letter/email will be sent to candidates latest the day after the kick-off meeting.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Volunteer Zoom Kick-Off Meeting&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
January 12, 2026 | 11:00 AM to 12:00 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Program Dates: January, 2026 –  April, 2026&#039;&#039;&#039; (13 weeks)&lt;br /&gt;
&lt;br /&gt;
Remote | Hybrid for GW employees and students (Ross Hall 5th floor)&lt;br /&gt;
&lt;br /&gt;
[[Volunteership Fall 2025|Fall 2025 Volunteership]] (Closed)&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteer Expectations ===&lt;br /&gt;
&lt;br /&gt;
# Minimum commitment of 10 hours per week.&lt;br /&gt;
# Progress updates via Slack at least 3 days per week (scrum).&lt;br /&gt;
# 30-minute Zoom meetings (during regular work hours) once a week or every other week with the assigned project point of contact (POC).&lt;br /&gt;
# Attend some lectures or seminars remotely (max 4-5).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;Important:&#039;&#039;&#039; If the scrum is not updated for 2 consecutive working days, the candidate will be automatically dropped from the program.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Potential Projects ===&lt;br /&gt;
We are excited to continue our bioinformatics volunteership program in Fall 2025. This program offers students the opportunity to work on bioinformatics projects supported by agencies such as the NIH, ARPA-H, and FDA. Participants will gain exposure to a variety of activities within a bioinformatics lab, including data analysis, computational biology, and genomics. If you are interested, please email mazumder_lab@gwu.edu your resume and a ranked list of the projects that interest you most. You can also indicate if you want to focus on specific areas that are of interest to you.&lt;br /&gt;
&lt;br /&gt;
# BiomarkerKB (biomarkerkb.org) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&lt;br /&gt;
# GlyGen (glygen.org) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information.&lt;br /&gt;
# ARGOS (argosdb.org) project: Analyze genomics data using HIVE to identify reference genome assemblies.&lt;br /&gt;
# PredictMod (hivelab.biochemistry.gwu.edu/predictmod) project. Curating PMIDs for intervention outcome prediction dataset LLM recommendation training.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen. &amp;lt;u&amp;gt;We are also looking for individuals who have previously worked with us to take on a coordinator role&amp;lt;/u&amp;gt;.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
==== 1. BiomarkerKB Biocuration Project Ideas ====&lt;br /&gt;
POC: Maria Kim, Cyrus Yeung, Jeet Vora&lt;br /&gt;
&lt;br /&gt;
# Curate biomarkers for a specific disease or for a treatment&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on NLP/LLM methods.&lt;br /&gt;
# Continue working on LLM methods started by volunteers in Fall 2025.&lt;br /&gt;
::: The data is available as well as some preliminary research and work done by previous volunteers in this area.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning (ML) Modeling Project ====&lt;br /&gt;
POC: Lori Krammer, Pat McNeely (optional)&lt;br /&gt;
&lt;br /&gt;
Volunteers will conduct ML modeling using publicly-available -omics datasets that were previously identified (see [[Recommended Publications for Intervention Outcome Prediction Models|https://hivelab.biochemistry.gwu.edu/wiki/Recommended_Publications_for_Intervention_Outcome_Prediction_Models]]). This volunteership will involve data harmonization, model training, and pipeline documentation.&lt;br /&gt;
&lt;br /&gt;
Tasks associated with this project include:&lt;br /&gt;
&lt;br /&gt;
# Exploring and understanding the data found in relevant PMIDs that can be used to train intervention outcome prediction models.&lt;br /&gt;
# Preparing the data for model training and model performance evaluation&lt;br /&gt;
# Testing the modeling tutorial, PredictMod platform, and associated project tools&lt;br /&gt;
# Documentation of the ML pipeline and testing results&lt;br /&gt;
Deliverables for this project include:&lt;br /&gt;
&lt;br /&gt;
# ML-ready datasets&lt;br /&gt;
# Trained model scripts&lt;br /&gt;
# Pipeline documentation captured in BioCompute Objects (BCOs) and testing reports&lt;br /&gt;
# Volunteership documentation (final report or weekly progress reports)&lt;br /&gt;
&lt;br /&gt;
Interested individuals should reach out to lorikrammer@gwu.edu. Please note that this project requires attendance at biweekly meetings and a final presentation of your work.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note:&#039;&#039; For anyone interested in ARGOS, you may be assigned to another project of your choice. This project is contingent on a contract extension. Please complete your project selection in order of preference.&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside&lt;br /&gt;
&lt;br /&gt;
Qualifications: basic/medium programming skills, knowledgeable of basic bioinformatics platforms and skills.&lt;br /&gt;
&lt;br /&gt;
# Curate and report on currently circulating pathogens to upload to ARGOS&lt;br /&gt;
## The student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found.&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
# QC Analysis using HIVE&lt;br /&gt;
## Analyze the curated pathogens using our QC ARGOS one-click pipeline.&lt;br /&gt;
## The results will be added to our ARGOS database.&lt;br /&gt;
# Report Results&lt;br /&gt;
## Defend your pathogens you have selected to be added to the database. Explain their importance and what value they would hold to the scientific community if they were added.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the Spring.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Requirements for Completion ===&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; The following are mandatory. Failure to complete any will result in an incomplete volunteer record.&lt;br /&gt;
&lt;br /&gt;
==== Documentation ====&lt;br /&gt;
All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&lt;br /&gt;
&lt;br /&gt;
==== Written Report ====&lt;br /&gt;
Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&lt;br /&gt;
&lt;br /&gt;
==== Presentation &amp;amp; Slide Submission ====&lt;br /&gt;
Present your work last week of the 13-week period.&lt;br /&gt;
&lt;br /&gt;
Slides must be submitted to the POCs.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program. Additional recognition will be given to the top three volunteers with exceptional presentations at the end of the program.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteers (TBD) ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Name&lt;br /&gt;
!Project Assigned&lt;br /&gt;
!POC Assigned&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/diya-kamalabharathy-62557935a/ Diya Kamalabharathy*]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer; Urnisha Bhuiyan; Rene Ranzinger&lt;br /&gt;
|PredictMod; Glyco web development&lt;br /&gt;
|-&lt;br /&gt;
|Sampurna Chakravorty&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer&lt;br /&gt;
|PredictMod; ARGOS; BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Vishal Muthusekaran*&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Maria Kim; Cyrus Yeung; Jeet Vora&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/ashley-tien/ Ashley Tien*]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer&lt;br /&gt;
|PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/conner-cognata/ Conner Cognata]&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Maria Kim; Cyrus Yeung; Jeet Vora&lt;br /&gt;
|BiomarkerKB; PredictMod; GlyGen biocuration&lt;br /&gt;
|-&lt;br /&gt;
|Venya Gulati&lt;br /&gt;
|ARGOS&lt;br /&gt;
|Christie Woodside&lt;br /&gt;
|ARGOS; PredictMod; BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Isaac Kim&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod; GlyGen biocuration; ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|Yashitha Pobbareddy&lt;br /&gt;
|ARGOS&lt;br /&gt;
|Christie Woodside&lt;br /&gt;
|ARGOS; GlyGen biocuration; BiomarkerKB&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;nowiki&amp;gt;*&amp;lt;/nowiki&amp;gt;Returning volunteer.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=1134</id>
		<title>Volunteership Fall 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=1134"/>
		<updated>2026-01-14T16:43:08Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* 2025 Volunteer Program Details */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== 2025 Volunteer Program Details ==&lt;br /&gt;
Click here for Spring 2026 Volunteership: [https://hivelab.biochemistry.gwu.edu/wiki/Volunteership_Spring_2026 Spring 2026 Volunteership]&lt;br /&gt;
&#039;&#039;&#039;Previous Volunteerships&#039;&#039;&#039;&lt;br /&gt;
[https://hivelab.biochemistry.gwu.edu/wiki/Volunteership_2025 Summer 2025 Volunteership] (Closed)&lt;br /&gt;
&lt;br /&gt;
=== Dates ===&lt;br /&gt;
&#039;&#039;&#039;Application Deadline&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 22, 2025, Noon (email your updated resume and projects in order of preference). Acceptance letter/email will be sent to candidates latest the day after the kick-off meeting.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Volunteer Zoom Kick-Off Meeting&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 25, 2025 | 4:00 to 5:00 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Program Dates: September 1st, 2025 – November 30th, 2025&#039;&#039;&#039; (13 weeks)&lt;br /&gt;
&lt;br /&gt;
Remote | Hybrid for GW employees and students (Ross Hall 5th floor)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteer Expectations ===&lt;br /&gt;
&lt;br /&gt;
# Minimum commitment of 10 hours per week.&lt;br /&gt;
# Progress updates via Slack at least 3 days per week (scrum).&lt;br /&gt;
# Regular Zoom meetings with the assigned project point of contact.&lt;br /&gt;
# Attend some lectures or seminars remotely (max 4-5).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;Important:&#039;&#039;&#039; If the scrum is not updated for 2 consecutive working days, the candidate will be automatically dropped from the program.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Potential Projects ===&lt;br /&gt;
We are excited to continue our bioinformatics volunteership program in Fall 2025. This program offers students the opportunity to work on bioinformatics projects supported by agencies such as the NIH, ARPA-H, and FDA. Participants will gain exposure to a variety of activities within a bioinformatics lab, including data analysis, computational biology, and genomics. If you are interested, please email mazumder_lab@gwu.edu your resume and a ranked list of the projects that interest you most. You can also indicate if you want to focus on specific areas that are of interest to you.&lt;br /&gt;
# BiomarkerKB (biomarkerkb.org) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&lt;br /&gt;
# GlyGen (glygen.org) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information.&lt;br /&gt;
# ARGOS (argosdb.org) project: Analyze genomics data using HIVE to identify reference genome assemblies.&lt;br /&gt;
# PredictMod (hivelab.biochemistry.gwu.edu/predictmod) project. Curating PMIDs for intervention outcome prediction dataset LLM recommendation training.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
==== 1. BiomarkerKB Biocuration Project Ideas ====&lt;br /&gt;
POC: Daniall Masood, Maria Kim&lt;br /&gt;
&lt;br /&gt;
# Curate biomarkers for a specific disease&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
# Continue working on LLM methods started by volunteers over the summer.&lt;br /&gt;
## The data is available as well as some preliminary research and work done by previous volunteers in this area.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer, Tianyi Wang, Pat McNeely (optional)&lt;br /&gt;
&lt;br /&gt;
Identifying relevant and useful publicly-available datasets for machine learning is currently a resource-intensive task. This curation project aims to develop a corpus for training an AI model to recommend PMIDs with publicly-available datasets useful for intervention outcome prediction models. The corpus will include an annotation spreadsheet + annotated PDFs for PubMed articles relevant to prostate, lung, breast cancers, biomarkers and glycans, and focus on indicators such as condition, intervention, and response.&lt;br /&gt;
&lt;br /&gt;
PMID curation involves:&lt;br /&gt;
&lt;br /&gt;
# Identify potentially relevant PMIDs that may have publicly-available datasets for training intervention outcome prediction models.&lt;br /&gt;
# Curate indicators of useful ML publications that could be used to train an LLM to recommend relevant publications for cancer modeling.&lt;br /&gt;
# Review peer curations and resolve annotation conflicts.&lt;br /&gt;
# Prepare a Wikipage to showcase the validated PMIDs.&lt;br /&gt;
&lt;br /&gt;
Interested individuals should reach out to lorikrammer@gwu.edu.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
# Update data tables for more efficient computations&lt;br /&gt;
## Student would review and input additional data and IDs in the tables/sheets used to perform computations. This would be manual work (but super important), but would require high attention to detail.&lt;br /&gt;
## Additional Work: Requires Python/shell coding background. Student would run scripts that prepare and format data tables that are pushed to data.argosdb.org. Coding knowledge is needed in case of errors, bugs, or other mishaps in the code. Ongoing work as computations are performed.&lt;br /&gt;
# Curate and report on current pathogens to upload to ARGOS&lt;br /&gt;
## Student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found.&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
# QC Analysis using HIVE&lt;br /&gt;
## Analyze the curated pathogens using our QC ARGOS one-click pipeline.&lt;br /&gt;
## The results will be added to our ARGOS database.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the Fall.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Requirements for Completion ===&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; The following are mandatory. Failure to complete any will result in an incomplete volunteer record.&lt;br /&gt;
&lt;br /&gt;
==== Documentation ====&lt;br /&gt;
All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&lt;br /&gt;
&lt;br /&gt;
==== Written Report ====&lt;br /&gt;
Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&lt;br /&gt;
&lt;br /&gt;
==== Presentation &amp;amp; Slide Submission ====&lt;br /&gt;
Present your work last week of the 13-week period.&lt;br /&gt;
&lt;br /&gt;
Slides must be submitted to the POCs.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteers ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Name&lt;br /&gt;
!Project Assigned&lt;br /&gt;
!POC Assigned&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/diya-kamalabharathy-62557935a/ Diya Kamalabharathy*]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod; Glyco web development&lt;br /&gt;
|-&lt;br /&gt;
|Harivinay P. Gujjula*&lt;br /&gt;
|GlyGen&lt;br /&gt;
|Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
|GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|Sparsh Gupta*&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Vishal Muthusekaran&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Miao Wang*&lt;br /&gt;
|ARGOS&lt;br /&gt;
|Christie Woodside, Jonathon Keeney&lt;br /&gt;
|ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|Anika Sikka&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/farah-kamila/ Farah Kamila]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod, ARGOS, BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/arhamur-rauf-2a61b3156 Arhamur Rauf]&lt;br /&gt;
|ARGOS&lt;br /&gt;
|Christie Woodside, Jonathon Keeney&lt;br /&gt;
|ARGOS, GlyGen, PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/ashley-tien/ Ashley Tien]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|ARGOS, PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|Namrata Oruganti&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;nowiki&amp;gt;*&amp;lt;/nowiki&amp;gt;Returning volunteer.&lt;br /&gt;
&lt;br /&gt;
== Fall Symposium ==&lt;br /&gt;
The Fall symposium will be held virtually.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Date:&#039;&#039;&#039; Nov 26th, 2025 (Wednesday)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Time:&#039;&#039;&#039; 3 - 5 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Zoom Link&#039;&#039;&#039; - https://gwu-edu.zoom.us/j/96518488501?jst=2&lt;br /&gt;
&lt;br /&gt;
=== Agenda (All times are in Eastern Standard Time) ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Time&lt;br /&gt;
!Project&lt;br /&gt;
!Presentation Title&lt;br /&gt;
!Presenter(s)&lt;br /&gt;
|-&lt;br /&gt;
|3:00-3:10 PM&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |Welcome &amp;amp; Introduction&lt;br /&gt;
|Raja Mazumder&lt;br /&gt;
|-&lt;br /&gt;
|3:10-3:35 PM&lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (Tianyi &amp;amp; Lori) intro&lt;br /&gt;
* 15 mins - PredictMod: PMID Curation for Intervention Outcome Prediction Models (IOPMs)&lt;br /&gt;
* 5 min QA&lt;br /&gt;
|Diya Kamalabharathy; Anika Sikka; Ashley Tien; Farah Kamila&lt;br /&gt;
|-&lt;br /&gt;
|3:35-4:00 PM&lt;br /&gt;
|GlyGen&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC intro (Urnisha, Rene, Kate)&lt;br /&gt;
* 15 mins - Curation of species metadata using LLM &amp;amp; Visualizing glycomics databases and their features&lt;br /&gt;
* 5 min QA&lt;br /&gt;
|Diya Kamalabharathy; Harivinay P. Gujjula&lt;br /&gt;
|-&lt;br /&gt;
|4:00-4:25 PM&lt;br /&gt;
|ARGOS&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (Christie) intro&lt;br /&gt;
* 15 mins -Curation of Pathogens and QC Analysis for the Argos Project QC analysis, representative genome selection Curation of genomes 1 &amp;amp; 2&lt;br /&gt;
* 5 mins QA&lt;br /&gt;
|Miao Wang; Arhamur Rauf&lt;br /&gt;
|-&lt;br /&gt;
|4:25-4:50 PM&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (Daniall &amp;amp; Maria) intro &lt;br /&gt;
* 15 mins - Leveraging Large Language Models to collect Biomarker data from PubMed Abstracts&lt;br /&gt;
* 5 mins QA&lt;br /&gt;
|Namrata Oruganti; Vishal Muthusekaran; Sparsh Gupta&lt;br /&gt;
|-&lt;br /&gt;
|4:50-5:00 PM&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |Remarks&lt;br /&gt;
|Raja Mazumder&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=1099</id>
		<title>Volunteership Fall 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=1099"/>
		<updated>2025-11-26T20:22:00Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Volunteers */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== 2025 Volunteer Program Details ==&lt;br /&gt;
&lt;br /&gt;
=== Dates ===&lt;br /&gt;
&#039;&#039;&#039;Application Deadline&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 22, 2025, Noon (email your updated resume and projects in order of preference). Acceptance letter/email will be sent to candidates latest the day after the kick-off meeting.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Volunteer Zoom Kick-Off Meeting&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 25, 2025 | 4:00 to 5:00 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Program Dates: September 1st, 2025 – November 30th, 2025&#039;&#039;&#039; (13 weeks)&lt;br /&gt;
&lt;br /&gt;
Remote | Hybrid for GW employees and students (Ross Hall 5th floor)&lt;br /&gt;
&lt;br /&gt;
[https://hivelab.biochemistry.gwu.edu/wiki/Volunteership_2025 Summer 2025 Volunteership] (Closed)&lt;br /&gt;
&lt;br /&gt;
[https://hivelab.biochemistry.gwu.edu/wiki/Volunteership_Spring_2026 Spring 2026 Volunteership]&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteer Expectations ===&lt;br /&gt;
&lt;br /&gt;
# Minimum commitment of 10 hours per week.&lt;br /&gt;
# Progress updates via Slack at least 3 days per week (scrum).&lt;br /&gt;
# Regular Zoom meetings with the assigned project point of contact.&lt;br /&gt;
# Attend some lectures or seminars remotely (max 4-5).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;Important:&#039;&#039;&#039; If the scrum is not updated for 2 consecutive working days, the candidate will be automatically dropped from the program.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Potential Projects ===&lt;br /&gt;
We are excited to continue our bioinformatics volunteership program in Fall 2025. This program offers students the opportunity to work on bioinformatics projects supported by agencies such as the NIH, ARPA-H, and FDA. Participants will gain exposure to a variety of activities within a bioinformatics lab, including data analysis, computational biology, and genomics. If you are interested, please email mazumder_lab@gwu.edu your resume and a ranked list of the projects that interest you most. You can also indicate if you want to focus on specific areas that are of interest to you.&lt;br /&gt;
# BiomarkerKB (biomarkerkb.org) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&lt;br /&gt;
# GlyGen (glygen.org) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information.&lt;br /&gt;
# ARGOS (argosdb.org) project: Analyze genomics data using HIVE to identify reference genome assemblies.&lt;br /&gt;
# PredictMod (hivelab.biochemistry.gwu.edu/predictmod) project. Curating PMIDs for intervention outcome prediction dataset LLM recommendation training.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
==== 1. BiomarkerKB Biocuration Project Ideas ====&lt;br /&gt;
POC: Daniall Masood, Maria Kim&lt;br /&gt;
&lt;br /&gt;
# Curate biomarkers for a specific disease&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
# Continue working on LLM methods started by volunteers over the summer.&lt;br /&gt;
## The data is available as well as some preliminary research and work done by previous volunteers in this area.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer, Tianyi Wang, Pat McNeely (optional)&lt;br /&gt;
&lt;br /&gt;
Identifying relevant and useful publicly-available datasets for machine learning is currently a resource-intensive task. This curation project aims to develop a corpus for training an AI model to recommend PMIDs with publicly-available datasets useful for intervention outcome prediction models. The corpus will include an annotation spreadsheet + annotated PDFs for PubMed articles relevant to prostate, lung, breast cancers, biomarkers and glycans, and focus on indicators such as condition, intervention, and response.&lt;br /&gt;
&lt;br /&gt;
PMID curation involves:&lt;br /&gt;
&lt;br /&gt;
# Identify potentially relevant PMIDs that may have publicly-available datasets for training intervention outcome prediction models.&lt;br /&gt;
# Curate indicators of useful ML publications that could be used to train an LLM to recommend relevant publications for cancer modeling.&lt;br /&gt;
# Review peer curations and resolve annotation conflicts.&lt;br /&gt;
# Prepare a Wikipage to showcase the validated PMIDs.&lt;br /&gt;
&lt;br /&gt;
Interested individuals should reach out to lorikrammer@gwu.edu.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
# Update data tables for more efficient computations&lt;br /&gt;
## Student would review and input additional data and IDs in the tables/sheets used to perform computations. This would be manual work (but super important), but would require high attention to detail.&lt;br /&gt;
## Additional Work: Requires Python/shell coding background. Student would run scripts that prepare and format data tables that are pushed to data.argosdb.org. Coding knowledge is needed in case of errors, bugs, or other mishaps in the code. Ongoing work as computations are performed.&lt;br /&gt;
# Curate and report on current pathogens to upload to ARGOS&lt;br /&gt;
## Student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found.&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
# QC Analysis using HIVE&lt;br /&gt;
## Analyze the curated pathogens using our QC ARGOS one-click pipeline.&lt;br /&gt;
## The results will be added to our ARGOS database.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the Fall.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Requirements for Completion ===&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; The following are mandatory. Failure to complete any will result in an incomplete volunteer record.&lt;br /&gt;
&lt;br /&gt;
==== Documentation ====&lt;br /&gt;
All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&lt;br /&gt;
&lt;br /&gt;
==== Written Report ====&lt;br /&gt;
Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&lt;br /&gt;
&lt;br /&gt;
==== Presentation &amp;amp; Slide Submission ====&lt;br /&gt;
Present your work last week of the 13-week period.&lt;br /&gt;
&lt;br /&gt;
Slides must be submitted to the POCs.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteers ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Name&lt;br /&gt;
!Project Assigned&lt;br /&gt;
!POC Assigned&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/diya-kamalabharathy-62557935a/ Diya Kamalabharathy*]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod; Glyco web development&lt;br /&gt;
|-&lt;br /&gt;
|Harivinay P. Gujjula*&lt;br /&gt;
|GlyGen&lt;br /&gt;
|Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
|GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|Sparsh Gupta*&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Vishal Muthusekaran&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Miao Wang*&lt;br /&gt;
|ARGOS&lt;br /&gt;
|Christie Woodside, Jonathon Keeney&lt;br /&gt;
|ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|Anika Sikka&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/farah-kamila/ Farah Kamila]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod, ARGOS, BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/arhamur-rauf-2a61b3156 Arhamur Rauf]&lt;br /&gt;
|ARGOS&lt;br /&gt;
|Christie Woodside, Jonathon Keeney&lt;br /&gt;
|ARGOS, GlyGen, PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/ashley-tien/ Ashley Tien]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|ARGOS, PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|Namrata Oruganti&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;nowiki&amp;gt;*&amp;lt;/nowiki&amp;gt;Returning volunteer.&lt;br /&gt;
&lt;br /&gt;
== Fall Symposium ==&lt;br /&gt;
The Fall symposium will be held virtually.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Date:&#039;&#039;&#039; Nov 26th, 2025 (Wednesday)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Time:&#039;&#039;&#039; 3 - 5 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Zoom Link&#039;&#039;&#039; - https://gwu-edu.zoom.us/j/96518488501?jst=2&lt;br /&gt;
&lt;br /&gt;
=== Agenda (All times are in Eastern Standard Time) ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Time&lt;br /&gt;
!Project&lt;br /&gt;
!Presentation Title&lt;br /&gt;
!Presenter(s)&lt;br /&gt;
|-&lt;br /&gt;
|3:00-3:10 PM&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |Welcome &amp;amp; Introduction&lt;br /&gt;
|Raja Mazumder&lt;br /&gt;
|-&lt;br /&gt;
|3:10-3:35 PM&lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (Tianyi &amp;amp; Lori) intro&lt;br /&gt;
* 15 mins - PredictMod: PMID Curation for Intervention Outcome Prediction Models (IOPMs)&lt;br /&gt;
* 5 min QA&lt;br /&gt;
|Diya Kamalabharathy; Anika Sikka; Ashley Tien; Farah Kamila&lt;br /&gt;
|-&lt;br /&gt;
|3:35-4:00 PM&lt;br /&gt;
|GlyGen&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC intro (Urnisha, Rene, Kate)&lt;br /&gt;
* 15 mins - Curation of species metadata using LLM &amp;amp; Visualizing glycomics databases and their features&lt;br /&gt;
* 5 min QA&lt;br /&gt;
|Diya Kamalabharathy; Harivinay P. Gujjula&lt;br /&gt;
|-&lt;br /&gt;
|4:00-4:25 PM&lt;br /&gt;
|ARGOS&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (Christie) intro&lt;br /&gt;
* 15 mins -Curation of Pathogens and QC Analysis for the Argos Project QC analysis, representative genome selection Curation of genomes 1 &amp;amp; 2&lt;br /&gt;
* 5 mins QA&lt;br /&gt;
|Miao Wang; Arhamur Rauf&lt;br /&gt;
|-&lt;br /&gt;
|4:25-4:50 PM&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (Daniall &amp;amp; Maria) intro &lt;br /&gt;
* 15 mins - Leveraging Large Language Models to collect Biomarker data from PubMed Abstracts&lt;br /&gt;
* 5 mins QA&lt;br /&gt;
|Namrata Oruganti; Vishal Muthusekaran; Sparsh Gupta&lt;br /&gt;
|-&lt;br /&gt;
|4:50-5:00 PM&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |Remarks&lt;br /&gt;
|Raja Mazumder&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=1095</id>
		<title>Volunteership Fall 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=1095"/>
		<updated>2025-11-26T17:52:10Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Agenda (All times are in Eastern Standard Time) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== 2025 Volunteer Program Details ==&lt;br /&gt;
&lt;br /&gt;
=== Dates ===&lt;br /&gt;
&#039;&#039;&#039;Application Deadline&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 22, 2025, Noon (email your updated resume and projects in order of preference). Acceptance letter/email will be sent to candidates latest the day after the kick-off meeting.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Volunteer Zoom Kick-Off Meeting&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 25, 2025 | 4:00 to 5:00 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Program Dates: September 1st, 2025 – November 30th, 2025&#039;&#039;&#039; (13 weeks)&lt;br /&gt;
&lt;br /&gt;
Remote | Hybrid for GW employees and students (Ross Hall 5th floor)&lt;br /&gt;
&lt;br /&gt;
[https://hivelab.biochemistry.gwu.edu/wiki/Volunteership_2025 Summer 2025 Volunteership] (Closed)&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteer Expectations ===&lt;br /&gt;
&lt;br /&gt;
# Minimum commitment of 10 hours per week.&lt;br /&gt;
# Progress updates via Slack at least 3 days per week (scrum).&lt;br /&gt;
# Regular Zoom meetings with the assigned project point of contact.&lt;br /&gt;
# Attend some lectures or seminars remotely (max 4-5).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;Important:&#039;&#039;&#039; If the scrum is not updated for 2 consecutive working days, the candidate will be automatically dropped from the program.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Potential Projects ===&lt;br /&gt;
We are excited to continue our bioinformatics volunteership program in Fall 2025. This program offers students the opportunity to work on bioinformatics projects supported by agencies such as the NIH, ARPA-H, and FDA. Participants will gain exposure to a variety of activities within a bioinformatics lab, including data analysis, computational biology, and genomics. If you are interested, please email mazumder_lab@gwu.edu your resume and a ranked list of the projects that interest you most. You can also indicate if you want to focus on specific areas that are of interest to you.&lt;br /&gt;
# BiomarkerKB (biomarkerkb.org) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&lt;br /&gt;
# GlyGen (glygen.org) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information.&lt;br /&gt;
# ARGOS (argosdb.org) project: Analyze genomics data using HIVE to identify reference genome assemblies.&lt;br /&gt;
# PredictMod (hivelab.biochemistry.gwu.edu/predictmod) project. Curating PMIDs for intervention outcome prediction dataset LLM recommendation training.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
==== 1. BiomarkerKB Biocuration Project Ideas ====&lt;br /&gt;
POC: Daniall Masood, Maria Kim&lt;br /&gt;
&lt;br /&gt;
# Curate biomarkers for a specific disease&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
# Continue working on LLM methods started by volunteers over the summer.&lt;br /&gt;
## The data is available as well as some preliminary research and work done by previous volunteers in this area.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer, Tianyi Wang, Pat McNeely (optional)&lt;br /&gt;
&lt;br /&gt;
Identifying relevant and useful publicly-available datasets for machine learning is currently a resource-intensive task. This curation project aims to develop a corpus for training an AI model to recommend PMIDs with publicly-available datasets useful for intervention outcome prediction models. The corpus will include an annotation spreadsheet + annotated PDFs for PubMed articles relevant to prostate, lung, breast cancers, biomarkers and glycans, and focus on indicators such as condition, intervention, and response.&lt;br /&gt;
&lt;br /&gt;
PMID curation involves:&lt;br /&gt;
&lt;br /&gt;
# Identify potentially relevant PMIDs that may have publicly-available datasets for training intervention outcome prediction models.&lt;br /&gt;
# Curate indicators of useful ML publications that could be used to train an LLM to recommend relevant publications for cancer modeling.&lt;br /&gt;
# Review peer curations and resolve annotation conflicts.&lt;br /&gt;
# Prepare a Wikipage to showcase the validated PMIDs.&lt;br /&gt;
&lt;br /&gt;
Interested individuals should reach out to lorikrammer@gwu.edu.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
# Update data tables for more efficient computations&lt;br /&gt;
## Student would review and input additional data and IDs in the tables/sheets used to perform computations. This would be manual work (but super important), but would require high attention to detail.&lt;br /&gt;
## Additional Work: Requires Python/shell coding background. Student would run scripts that prepare and format data tables that are pushed to data.argosdb.org. Coding knowledge is needed in case of errors, bugs, or other mishaps in the code. Ongoing work as computations are performed.&lt;br /&gt;
# Curate and report on current pathogens to upload to ARGOS&lt;br /&gt;
## Student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found.&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
# QC Analysis using HIVE&lt;br /&gt;
## Analyze the curated pathogens using our QC ARGOS one-click pipeline.&lt;br /&gt;
## The results will be added to our ARGOS database.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the Fall.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Requirements for Completion ===&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; The following are mandatory. Failure to complete any will result in an incomplete volunteer record.&lt;br /&gt;
&lt;br /&gt;
==== Documentation ====&lt;br /&gt;
All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&lt;br /&gt;
&lt;br /&gt;
==== Written Report ====&lt;br /&gt;
Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&lt;br /&gt;
&lt;br /&gt;
==== Presentation &amp;amp; Slide Submission ====&lt;br /&gt;
Present your work last week of the 13-week period.&lt;br /&gt;
&lt;br /&gt;
Slides must be submitted to the POCs.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteers ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Name&lt;br /&gt;
!Project Assigned&lt;br /&gt;
!POC Assigned&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/diya-kamalabharathy-62557935a/ Diya Kamalabharathy*]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod; Glyco web development&lt;br /&gt;
|-&lt;br /&gt;
|Harivinay P. Gujjula*&lt;br /&gt;
|GlyGen&lt;br /&gt;
|Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
|GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|Sparsh Gupta*&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Vishal Muthusekaran&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Miao Wang*&lt;br /&gt;
|ARGOS&lt;br /&gt;
|Christie Woodside, Jonathon Keeney&lt;br /&gt;
|ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|Anika Sikka&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/farah-kamila/ Farah Kamila]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|PredictMod, ARGOS, BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Arhamur Rauf&lt;br /&gt;
|ARGOS&lt;br /&gt;
|Christie Woodside, Jonathon Keeney&lt;br /&gt;
|ARGOS, GlyGen, PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/ashley-tien/ Ashley Tien]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|Lori Krammer, Tianyi Wang&lt;br /&gt;
|ARGOS, PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|Namrata Oruganti&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|Daniall Masood, Maria Kim&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;nowiki&amp;gt;*&amp;lt;/nowiki&amp;gt;Returning volunteer.&lt;br /&gt;
&lt;br /&gt;
== Fall Symposium ==&lt;br /&gt;
The Fall symposium will be held virtually.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Date:&#039;&#039;&#039; Nov 26th, 2025 (Wednesday)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Time:&#039;&#039;&#039; 3 - 5 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Zoom Link&#039;&#039;&#039; - https://gwu-edu.zoom.us/j/96518488501?jst=2&lt;br /&gt;
&lt;br /&gt;
=== Agenda (All times are in Eastern Standard Time) ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Time&lt;br /&gt;
!Project&lt;br /&gt;
!Presentation Title&lt;br /&gt;
!Presenter(s)&lt;br /&gt;
|-&lt;br /&gt;
|3:00-3:10 PM&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |Welcome &amp;amp; Introduction&lt;br /&gt;
|Raja Mazumder&lt;br /&gt;
|-&lt;br /&gt;
|3:10-3:35 PM&lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (Tianyi &amp;amp; Lori) intro&lt;br /&gt;
* 15 mins - PredictMod: PMID Curation for Intervention Outcome Prediction Models (IOPMs)&lt;br /&gt;
* 5 min QA&lt;br /&gt;
|Diya Kamalabharathy; Anika Sikka; Ashley Tien; Farah Kamila&lt;br /&gt;
|-&lt;br /&gt;
|3:35-4:00 PM&lt;br /&gt;
|GlyGen&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC intro&lt;br /&gt;
* 15 mins - Curation of species metadata using LLM &amp;amp; Visualizing glycomics databases and their features&lt;br /&gt;
* 5 min QA&lt;br /&gt;
|Diya Kamalabharathy; Harivinay P. Gujjula&lt;br /&gt;
|-&lt;br /&gt;
|4:00-4:25 PM&lt;br /&gt;
|ARGOS&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (christie) intro&lt;br /&gt;
* 15 mins -Curation of Pathogens and QC Analysis for the Argos Project QC analysis, representative genome selection Curation of genomes 1 &amp;amp; 2&lt;br /&gt;
* 5 mins QA&lt;br /&gt;
|Miao Wang; Arhamur Rauf&lt;br /&gt;
|-&lt;br /&gt;
|4:25-4:50 PM&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
* 5 min POC (Daniall and Maria) intro &lt;br /&gt;
* 15 mins - Leveraging Large Language Models to collect Biomarker data from PubMed Abstracts&lt;br /&gt;
* 5 mins QA&lt;br /&gt;
|Namrata Oruganti; Vishal Muthusekaran; Sparsh Gupta&lt;br /&gt;
|-&lt;br /&gt;
|4:50-5:00 PM&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |Remarks&lt;br /&gt;
|Raja Mazumder&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Spring_2026&amp;diff=1089</id>
		<title>Volunteership Spring 2026</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Spring_2026&amp;diff=1089"/>
		<updated>2025-11-14T21:16:17Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* 4. PredictMod Machine Learning Project Ideas */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== 2026 Spring Volunteer Program Details ==&lt;br /&gt;
&lt;br /&gt;
=== Dates ===&lt;br /&gt;
&#039;&#039;&#039;Application Deadline&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
January 9, 2026, Noon (email your updated resume and projects in order of preference). Acceptance letter/email will be sent to candidates latest the day after the kick-off meeting.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Volunteer Zoom Kick-Off Meeting&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
January 12, 2026 | 4:00 to 5:00 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Program Dates: January, 2026 –  April, 2026&#039;&#039;&#039; (13 weeks)&lt;br /&gt;
&lt;br /&gt;
Remote | Hybrid for GW employees and students (Ross Hall 5th floor)&lt;br /&gt;
&lt;br /&gt;
[[Volunteership Fall 2025|Fall 2025 Volunteership]] (Closed)&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteer Expectations ===&lt;br /&gt;
&lt;br /&gt;
# Minimum commitment of 10 hours per week.&lt;br /&gt;
# Progress updates via Slack at least 3 days per week (scrum).&lt;br /&gt;
# Regular Zoom meetings with the assigned project point of contact.&lt;br /&gt;
# Attend some lectures or seminars remotely (max 4-5).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;Important:&#039;&#039;&#039; If the scrum is not updated for 2 consecutive working days, the candidate will be automatically dropped from the program.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Potential Projects ===&lt;br /&gt;
We are excited to continue our bioinformatics volunteership program in Fall 2025. This program offers students the opportunity to work on bioinformatics projects supported by agencies such as the NIH, ARPA-H, and FDA. Participants will gain exposure to a variety of activities within a bioinformatics lab, including data analysis, computational biology, and genomics. If you are interested, please email mazumder_lab@gwu.edu your resume and a ranked list of the projects that interest you most. You can also indicate if you want to focus on specific areas that are of interest to you.&lt;br /&gt;
&lt;br /&gt;
# BiomarkerKB (biomarkerkb.org) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&lt;br /&gt;
# GlyGen (glygen.org) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information.&lt;br /&gt;
# ARGOS (argosdb.org) project: Analyze genomics data using HIVE to identify reference genome assemblies.&lt;br /&gt;
# PredictMod (hivelab.biochemistry.gwu.edu/predictmod) project. Curating PMIDs for intervention outcome prediction dataset LLM recommendation training.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
==== 1. BiomarkerKB Biocuration Project Ideas ====&lt;br /&gt;
POC: Daniall Masood, Maria Kim&lt;br /&gt;
&lt;br /&gt;
# Curate biomarkers for a specific disease&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
# Continue working on LLM methods started by volunteers over the summer.&lt;br /&gt;
## The data is available as well as some preliminary research and work done by previous volunteers in this area.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer, Tianyi Wang, Pat McNeely (optional)&lt;br /&gt;
&lt;br /&gt;
Identifying relevant and useful publicly-available datasets for machine learning is currently a resource-intensive task. This curation project aims to develop a corpus for training an AI model to recommend PMIDs with publicly-available datasets useful for intervention outcome prediction models. The corpus will include an annotation spreadsheet + annotated PDFs for PubMed articles relevant to prostate, lung, breast cancers, biomarkers and glycans, and focus on indicators such as condition, intervention, and response.&lt;br /&gt;
&lt;br /&gt;
PMID curation involves:&lt;br /&gt;
&lt;br /&gt;
# Identify potentially relevant PMIDs that may have publicly-available datasets for training intervention outcome prediction models.&lt;br /&gt;
# Curate indicators of useful ML publications that could be used to train an LLM to recommend relevant publications for cancer modeling.&lt;br /&gt;
# Review peer curations and resolve annotation conflicts.&lt;br /&gt;
# Prepare a Wikipage to showcase the validated PMIDs.&lt;br /&gt;
&lt;br /&gt;
Interested individuals should reach out to lorikrammer@gwu.edu. Please note that this project requires attendance at biweekly meetings and weekly 1-2 paragraph reports.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note:&#039;&#039; For anyone interested in ARGOS, you may be assigned to another project of your choice. This project is contingent on a contract extension. Please complete your project selection in order of preference.&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
Qualifications: basic/medium programming skills, knowledgeable of basic bioinformatics platforms and skills.&lt;br /&gt;
&lt;br /&gt;
# Curate and report on currently circulating pathogens to upload to ARGOS&lt;br /&gt;
## The student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found.&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
# QC Analysis using HIVE&lt;br /&gt;
## Analyze the curated pathogens using our QC ARGOS one-click pipeline.&lt;br /&gt;
## The results will be added to our ARGOS database.&lt;br /&gt;
# Report Results&lt;br /&gt;
## Defend your pathogens you have selected to be added to the database. Explain their importance and what value they would hold to the scientific community if they were added.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the Spring.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Requirements for Completion ===&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; The following are mandatory. Failure to complete any will result in an incomplete volunteer record.&lt;br /&gt;
&lt;br /&gt;
==== Documentation ====&lt;br /&gt;
All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&lt;br /&gt;
&lt;br /&gt;
==== Written Report ====&lt;br /&gt;
Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&lt;br /&gt;
&lt;br /&gt;
==== Presentation &amp;amp; Slide Submission ====&lt;br /&gt;
Present your work last week of the 13-week period.&lt;br /&gt;
&lt;br /&gt;
Slides must be submitted to the POCs.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteers (TBD) ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Name&lt;br /&gt;
!Project Assigned&lt;br /&gt;
!POC Assigned&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;nowiki&amp;gt;*&amp;lt;/nowiki&amp;gt;Returning volunteer.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_FAQs&amp;diff=1052</id>
		<title>FDA-ARGOS FAQs</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_FAQs&amp;diff=1052"/>
		<updated>2025-09-19T17:27:42Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
Back to [[FDA-ARGOS WIKI|Home Page]] for FDA-ARGOS&lt;br /&gt;
&lt;br /&gt;
===== What is the ArgosDB and how is it organized? =====&lt;br /&gt;
ArgosDB was developed as a result of expanded funding for the FDA-ARGOS project, which is described in detail in the [[About Argos DataBase|About]] page of this website. The database stores cross-kingdom QC attributes of clinically relevant organisms organized into respective datasets. The current datasets (as of September 2025) are: ngsQC_ARGOS, ngsQC_ARGOS_extended, assemblyQC_ARGOS, assemblyQC_ARGOS_extended, biosampleMeta_ARGOS, and biosampleMeta_ARGOS_extended. These datasets are associated with core QC protocols, which are documented via Biocompute Objects (BCO) and organized under their BCO IDs.  &lt;br /&gt;
&lt;br /&gt;
When new QC data has been produced for an organism of interest, the respective dataset(s) is/are appended, which is documented per data release in the FDA-ARGOS GitHub (&amp;lt;nowiki&amp;gt;https://github.com/FDA-ARGOS&amp;lt;/nowiki&amp;gt;) and on the database itself. All datasets are in alignment with the current data dictionary (v1.6.1 as of September 2025), which guides the QC process for that dataset as well as the column headers for a given dataset. Datasets are available as either .tsv or FASTA files if associated with a genome assembly, and all datasets and BCOs are available for download. All data provenance and curation are captured and reproducible via their BCO. Additional datasets include data from the original FDA BioProject, the Data Dictionary, Drug Resistance Mutations, Genome Assemblies (multiple), and a mapping key that assists in linking all the available data via important accessions. &lt;br /&gt;
&lt;br /&gt;
A total of 14 datasets are available as of 09/2025.&lt;br /&gt;
&lt;br /&gt;
===== How can I view or access the previous versions of the data? =====&lt;br /&gt;
First, go to the Release History tab on the ARGOS home page. Next, click &#039;details&#039; on the desired data object. Then select the desired version and data from the version transition dropbox on the top left corner and view metrics such as field count, fields added, fields removed, row count, row count prev, rows count change, ID count, IDs added, and IDs removed.&lt;br /&gt;
&lt;br /&gt;
===== What does the schema version in the datasets refer to? =====&lt;br /&gt;
The schema relates to the organization of the ARGOS data within the data model. The version is reflective of the FDA-ARGOS data dictionary version that is currently applied to all updated datasets. As of September 2025, the current schema is v1.6.1 and can be found in the [https://github.com/FDA-ARGOS/data.argosdb/tree/main/data_dictionary/v1.6.1 FDA-ARGOS GitHub]. &lt;br /&gt;
&lt;br /&gt;
===== Does Argos have a tutorial on how to use the site? =====&lt;br /&gt;
Yes! Please follow the instructions below of how to navigate the DB.   &lt;br /&gt;
&lt;br /&gt;
====== How to find and search a dataset ======&lt;br /&gt;
On data.argosdb.org home page, you can search for a dataset by entering the keyword in Search Datasets. &lt;br /&gt;
&lt;br /&gt;
Keywords can be BCO ID, organism name, or even a term that describes biological processes. In the following example, three results appear upon the search for Ebola.  &lt;br /&gt;
To further narrow down the result, select filters on the left sidebar. Alternatively, users can search datasets by selecting relevant filters on the left sidebar.&lt;br /&gt;
[[File:Screenshot 2025-09-19 at 1.16.41 PM.png|none|thumb|719x719px|data.argosdb Home Page]]&lt;br /&gt;
[[File:Screenshot 2025-09-19 at 1.17.57 PM.png|none|thumb|722x722px|&#039;Ebola&#039; searched in the search bar of the ARGOS database]]&lt;br /&gt;
&lt;br /&gt;
====== How to select a dataset ======&lt;br /&gt;
Next, to select a dataset, click on view details under DETAILS. Previous released versioned datasets are available upon clicking the dropdown button &lt;br /&gt;
[[File:Dataset.png|thumb|722x722px|view of an example dataset after clicking on the &#039;...view details&#039; link on the homepage. The dropdown menu at the top lets you select data versions.|none]] &lt;br /&gt;
&lt;br /&gt;
====== How to view and download the BCO that is corresponding to its dataset ======&lt;br /&gt;
To download the dataset, click on the DOWNLOADS tab and select the download format for the target dataset. BCO JSON will be downloaded and automatically opened as a .txt file upon clicking on Download BCO. Dataset will be downloaded and automatically opened either as .tsv or .csv file upon clicking on Download dataset file.&lt;br /&gt;
[[File:Bco.png|thumb|720x720px|BCO JSON tab of the dataset.|none]]&lt;br /&gt;
[[File:Screenshot 2025-03-06 at 2.08.02 PM.png|thumb|724x724px|Downloads tab for the dataset. BCO and table can be downloaded here.|none]]&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_FAQs&amp;diff=1051</id>
		<title>FDA-ARGOS FAQs</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_FAQs&amp;diff=1051"/>
		<updated>2025-09-19T17:18:58Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
Back to [[FDA-ARGOS WIKI|Home Page]] for FDA-ARGOS&lt;br /&gt;
&lt;br /&gt;
===== What is the ArgosDB and how is it organized? =====&lt;br /&gt;
ArgosDB was developed as a result of expanded funding for the FDA-ARGOS project, which is described in detail in the [[About Argos DataBase|About]] page of this website. The database stores cross-kingdom QC attributes of clinically relevant organisms organized into respective datasets. The current datasets (as of September 2025) are: ngsQC_ARGOS, ngsQC_ARGOS_extended, assemblyQC_ARGOS, assemblyQC_ARGOS_extended, biosampleMeta_ARGOS, and biosampleMeta_ARGOS_extended. These datasets are associated with core QC protocols, which are documented via Biocompute Objects (BCO) and organized under their BCO IDs.  &lt;br /&gt;
&lt;br /&gt;
When new QC data has been produced for an organism of interest, the respective dataset(s) is/are appended, which is documented per data release in the FDA-ARGOS GitHub (&amp;lt;nowiki&amp;gt;https://github.com/FDA-ARGOS&amp;lt;/nowiki&amp;gt;) and on the database itself. All datasets are in alignment with the current data dictionary (v1.6.1 as of September 2025), which guides the QC process for that dataset as well as the column headers for a given dataset. Datasets are available as either .tsv or FASTA files if associated with a genome assembly, and all datasets and BCOs are available for download. All data provenance and curation are captured and reproducible via their BCO. Additional datasets include data from the original FDA BioProject, the Data Dictionary, Drug Resistance Mutations, Genome Assemblies (multiple), and a mapping key that assists in linking all the available data via important accessions. &lt;br /&gt;
&lt;br /&gt;
A total of 14 datasets are available as of 09/2025.&lt;br /&gt;
&lt;br /&gt;
===== How can I view or access the previous versions of the data? =====&lt;br /&gt;
First, go to the Release History tab on the ARGOS home page. Next, click &#039;details&#039; on the desired data object. Then select the desired version and data from the version transition dropbox on the top left corner and view metrics such as field count, fields added, fields removed, row count, row count prev, rows count change, ID count, IDs added, and IDs removed.&lt;br /&gt;
&lt;br /&gt;
===== What does the schema version in the datasets refer to? =====&lt;br /&gt;
The schema relates to the organization of the ARGOS data within the data model. The version is reflective of the FDA-ARGOS data dictionary version that is currently applied to all updated datasets. As of September 2025, the current schema is v1.6.1 and can be found in the [https://github.com/FDA-ARGOS/data.argosdb/tree/main/data_dictionary/v1.6 FDA-ARGOS GitHub]. &lt;br /&gt;
&lt;br /&gt;
===== Does Argos have a tutorial on how to use the site? =====&lt;br /&gt;
Yes! Please follow the instructions below of how to navigate the DB.   &lt;br /&gt;
&lt;br /&gt;
====== How to find and search a dataset ======&lt;br /&gt;
On data.argosdb.org home page, you can search for a dataset by entering the keyword in Search Datasets. &lt;br /&gt;
&lt;br /&gt;
Keywords can be BCO ID, organism name, or even a term that describes biological processes. In the following example, three results appear upon the search for Ebola.  &lt;br /&gt;
To further narrow down the result, select filters on the left sidebar. Alternatively, users can search datasets by selecting relevant filters on the left sidebar.&lt;br /&gt;
[[File:Screenshot 2025-09-19 at 1.16.41 PM.png|none|thumb|719x719px|data.argosdb Home Page]]&lt;br /&gt;
[[File:Screenshot 2025-09-19 at 1.17.57 PM.png|none|thumb|722x722px|&#039;Ebola&#039; searched in the search bar of the ARGOS database]]&lt;br /&gt;
&lt;br /&gt;
====== How to select a dataset ======&lt;br /&gt;
Next, to select a dataset, click on view details under DETAILS. Previous released versioned datasets are available upon clicking the dropdown button &lt;br /&gt;
[[File:Dataset.png|thumb|722x722px|view of an example dataset after clicking on the &#039;...view details&#039; link on the homepage. The dropdown menu at the top lets you select data versions.|none]] &lt;br /&gt;
&lt;br /&gt;
====== How to view and download the BCO that is corresponding to its dataset ======&lt;br /&gt;
To download the dataset, click on the DOWNLOADS tab and select the download format for the target dataset. BCO JSON will be downloaded and automatically opened as a .txt file upon clicking on Download BCO. Dataset will be downloaded and automatically opened either as .tsv or .csv file upon clicking on Download dataset file.&lt;br /&gt;
[[File:Bco.png|thumb|720x720px|BCO JSON tab of the dataset.|none]]&lt;br /&gt;
[[File:Screenshot 2025-03-06 at 2.08.02 PM.png|thumb|724x724px|Downloads tab for the dataset. BCO and table can be downloaded here.|none]]&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_FAQs&amp;diff=1050</id>
		<title>FDA-ARGOS FAQs</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_FAQs&amp;diff=1050"/>
		<updated>2025-09-19T17:18:41Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
Back to [[FDA-ARGOS WIKI|Home Page]] for FDA-ARGOS&lt;br /&gt;
&lt;br /&gt;
===== What is the ArgosDB and how is it organized? =====&lt;br /&gt;
ArgosDB was developed as a result of expanded funding for the FDA-ARGOS project, which is described in detail in the [[About Argos DataBase|About]] page of this website. The database stores cross-kingdom QC attributes of clinically relevant organisms organized into respective datasets. The current datasets (as of September 2025) are: ngsQC_ARGOS, ngsQC_ARGOS_extended, assemblyQC_ARGOS, assemblyQC_ARGOS_extended, biosampleMeta_ARGOS, and biosampleMeta_ARGOS_extended. These datasets are associated with core QC protocols, which are documented via Biocompute Objects (BCO) and organized under their BCO IDs.  &lt;br /&gt;
&lt;br /&gt;
When new QC data has been produced for an organism of interest, the respective dataset(s) is/are appended, which is documented per data release in the FDA-ARGOS GitHub (&amp;lt;nowiki&amp;gt;https://github.com/FDA-ARGOS&amp;lt;/nowiki&amp;gt;) and on the database itself. All datasets are in alignment with the current data dictionary (v1.6.1 as of September 2025), which guides the QC process for that dataset as well as the column headers for a given dataset. Datasets are available as either .tsv or FASTA files if associated with a genome assembly, and all datasets and BCOs are available for download. All data provenance and curation are captured and reproducible via their BCO. Additional datasets include data from the original FDA BioProject, the Data Dictionary, Drug Resistance Mutations, Genome Assemblies (multiple), and a mapping key that assists in linking all the available data via important accessions. &lt;br /&gt;
&lt;br /&gt;
A total of 14 datasets are available as of 09/2025.&lt;br /&gt;
&lt;br /&gt;
===== How can I view or access the previous versions of the data? =====&lt;br /&gt;
First, go to the Release History tab on the ARGOS home page. Next, click &#039;details&#039; on the desired data object. Then select the desired version and data from the version transition dropbox on the top left corner and view metrics such as field count, fields added, fields removed, row count, row count prev, rows count change, ID count, IDs added, and IDs removed.&lt;br /&gt;
&lt;br /&gt;
===== What does the schema version in the datasets refer to? =====&lt;br /&gt;
The schema relates to the organization of the ARGOS data within the data model. The version is reflective of the FDA-ARGOS data dictionary version that is currently applied to all updated datasets. As of September 2025, the current schema is v1.6.1 and can be found in the [https://github.com/FDA-ARGOS/data.argosdb/tree/main/data_dictionary/v1.6 FDA-ARGOs GitHub]. &lt;br /&gt;
&lt;br /&gt;
===== Does Argos have a tutorial on how to use the site? =====&lt;br /&gt;
Yes! Please follow the instructions below of how to navigate the DB.   &lt;br /&gt;
&lt;br /&gt;
====== How to find and search a dataset ======&lt;br /&gt;
On data.argosdb.org home page, you can search for a dataset by entering the keyword in Search Datasets. &lt;br /&gt;
&lt;br /&gt;
Keywords can be BCO ID, organism name, or even a term that describes biological processes. In the following example, three results appear upon the search for Ebola.  &lt;br /&gt;
To further narrow down the result, select filters on the left sidebar. Alternatively, users can search datasets by selecting relevant filters on the left sidebar.&lt;br /&gt;
[[File:Screenshot 2025-09-19 at 1.16.41 PM.png|none|thumb|719x719px|data.argosdb Home Page]]&lt;br /&gt;
[[File:Screenshot 2025-09-19 at 1.17.57 PM.png|none|thumb|722x722px|&#039;Ebola&#039; searched in the search bar of the ARGOS database]]&lt;br /&gt;
&lt;br /&gt;
====== How to select a dataset ======&lt;br /&gt;
Next, to select a dataset, click on view details under DETAILS. Previous released versioned datasets are available upon clicking the dropdown button &lt;br /&gt;
[[File:Dataset.png|thumb|722x722px|view of an example dataset after clicking on the &#039;...view details&#039; link on the homepage. The dropdown menu at the top lets you select data versions.|none]] &lt;br /&gt;
&lt;br /&gt;
====== How to view and download the BCO that is corresponding to its dataset ======&lt;br /&gt;
To download the dataset, click on the DOWNLOADS tab and select the download format for the target dataset. BCO JSON will be downloaded and automatically opened as a .txt file upon clicking on Download BCO. Dataset will be downloaded and automatically opened either as .tsv or .csv file upon clicking on Download dataset file.&lt;br /&gt;
[[File:Bco.png|thumb|720x720px|BCO JSON tab of the dataset.|none]]&lt;br /&gt;
[[File:Screenshot 2025-03-06 at 2.08.02 PM.png|thumb|724x724px|Downloads tab for the dataset. BCO and table can be downloaded here.|none]]&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_2025-09-19_at_1.17.57_PM.png&amp;diff=1049</id>
		<title>File:Screenshot 2025-09-19 at 1.17.57 PM.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_2025-09-19_at_1.17.57_PM.png&amp;diff=1049"/>
		<updated>2025-09-19T17:18:18Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;ebola search image sept 2025&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_2025-09-19_at_1.16.41_PM.png&amp;diff=1048</id>
		<title>File:Screenshot 2025-09-19 at 1.16.41 PM.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_2025-09-19_at_1.16.41_PM.png&amp;diff=1048"/>
		<updated>2025-09-19T17:17:07Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;data.argosdb Home Page 2025&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOS_Contact_Us&amp;diff=1047</id>
		<title>ARGOS Contact Us</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOS_Contact_Us&amp;diff=1047"/>
		<updated>2025-09-19T17:11:51Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;To contact us for any inquires, &lt;br /&gt;
* email mazumder_lab@gwu.edu&lt;br /&gt;
We would love feedback on our website, data, or suggestions on data to include. &lt;br /&gt;
&lt;br /&gt;
Feel free to reach out to us about the data published in data.argosdb.org or any questions you may have.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Back to [[FDA-ARGOS WIKI|Home Page]] for FDA-ARGOS.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=1046</id>
		<title>ARGOSQC Usage Tutorial</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=1046"/>
		<updated>2025-09-19T17:10:02Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
Back to [[FDA-ARGOS WIKI|Home Page]] for FDA-ARGOS&lt;br /&gt;
&lt;br /&gt;
HIVE3 one-click pipeline tutorial for the FDA HIVE instance. This protocol will guide the user in running single and batch-mode QC computations. HIVE3 is an instance of HIVE not owned by the FDA which can be directly modified by Vahan or others on our team with permissions.&lt;br /&gt;
&lt;br /&gt;
= Required User Information =&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Protocol Version&#039;&#039;&#039; &lt;br /&gt;
|1.0&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Instance&#039;&#039;&#039;&lt;br /&gt;
|3&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Link&#039;&#039;&#039;&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=login&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Overview =&lt;br /&gt;
We constructed a QC one-click pipeline that takes user specified organism information and combines the 3 core ARGOS workflows to produce 5 different result datasets in JSON format (Figure 1). 3 out of the 5 result JSONs (assemblyQC, ngsQC, and biosampleMeta) have been updated and added to data.argosdb.&lt;br /&gt;
&lt;br /&gt;
To register your account, navigate to the link under the “Required User Information”. At the top right there will be a tab saying “register”. Fill out the appropriate fields and submit. Once submitted, please email mazumder_lab@gwu.edu so we can verify your account.&lt;br /&gt;
&lt;br /&gt;
The ARGOS pipeline can be accessed via the dropdown menu ‘Projects’ at the upper right hand screen and then under ‘Argos’. The pipeline in HIVE3 is located in the “Required User Information” section in the beginning of this protocol. &lt;br /&gt;
&lt;br /&gt;
After a successful login, you will be navigated to the home page. Use the menu at the top right corner under projects to access the ARGOS pipeline or use this URL to access the ARGOS QC pipeline on HIVE3:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=argos-alqc&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[File:Screenshot figure 1.png|thumb|667x667px|&#039;&#039;&#039;Figure 1.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline. |none]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Input values:&#039;&#039; ==&lt;br /&gt;
On the ARGOS Pipeline input setting page, under the General tab, is where our data inputs for a single and batch computation will be taken. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Name:&#039;&#039;&#039; Give the computation a name. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Folder:&#039;&#039;&#039; Give the folder where your computations, data, and steps will be stored.&lt;br /&gt;
&lt;br /&gt;
* Can use: _ or -  or &amp;amp;  and letters + numbers&lt;br /&gt;
* Cannot use: / : ; , \ “ ” ‘ ’ &lt;br /&gt;
* Yet, you use / to create a sub folder, but that is not recommended. Manually moving the subfolder is best.&lt;br /&gt;
* Ex folder: Influenza A (h5n1)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reads:&#039;&#039;&#039; information needed for ngsQC &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;SRR&#039;&#039;&#039;: the SRR accession number, can be multiple per organism by using a “,” or populate extra fields by clicking on the gray + sign. This tool uses the NCBI SRA Fasterq function to grab the fastq files directly from NCBI without the user needing to import them to HIVE.  &lt;br /&gt;
* &#039;&#039;&#039;HIVE reads:&#039;&#039;&#039; Drop down menu can select reads already uploaded into HIVE, either from previous computations or manual uploads. &lt;br /&gt;
* See in Figure 2a&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the SRR input box, rather the HIVE IDs menu, but you can if you want to, it will just search within HIVE. See ngsQC Protocol for how to upload SRR information using external downloader. This external downloader process is the same as in HIVE2 and 1.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reference:&#039;&#039;&#039; Information used for the assemblyQC portion of the algorithm.&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Reference accession:&#039;&#039;&#039; This is the REFSEQ or Genbank accession number from NCBI or Genbank.&lt;br /&gt;
* &#039;&#039;&#039;Assembly ID&#039;&#039;&#039;: This is the ASSEMBLY accession number from NCBI. &lt;br /&gt;
* &#039;&#039;&#039;HIVE genome:&#039;&#039;&#039; Use the drop down menu to select a reference genome that has already been uploaded into HIVE. &lt;br /&gt;
* See in Figure 2b&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the assembly ID or reference accession input box, but you can if you want to. See AssemblyQC Protocol for how to upload assembly information using an external downloader or local upload. This process is the same as what is done in HIVE2.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Metadata: &#039;&#039;&#039; Used to grab information necessary to fill out the BiosampleMeta_HIVE document.&lt;br /&gt;
&lt;br /&gt;
* Biosample Accession: The optional accession number for the Biosample that was reported to be used when creating the assembly and will be linked to the SRR fastq files used for the ngsQC portion of the algorithm. This step is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Coding Table:&#039;&#039;&#039; Dropdown of genetic codon tables to be used for your computation, depending on the organism to be computed. The default is human, viral (Standard).&lt;br /&gt;
&lt;br /&gt;
* Tip: NCBI Taxonomy will list the codon table for each organism on their taxonomy page.&lt;br /&gt;
[[File:Screenshot figure 2 a.png|none|thumb|657x657px|&#039;&#039;&#039;Figure 2.&#039;&#039;&#039; The ARGOS_QC algorithm page input with all NCBI information for a test organism, Salmonella enterica.]]&lt;br /&gt;
[[File:Screenshot fig 2a.png|none|thumb|508x508px|&#039;&#039;&#039;Figure 2 a)&#039;&#039;&#039; The SRR accession field contains both SRR fastqs for the organism that correspond to the biosample.]]&lt;br /&gt;
[[File:Screenshot figure 2b.png|none|thumb|502x502px|&#039;&#039;&#039;Figure 2 b)&#039;&#039;&#039; The &#039;&#039;&#039;Reference Accession&#039;&#039;&#039; is the RefSeq Nucleotide accession number from NCBI, the &#039;&#039;&#039;Assembly ID&#039;&#039;&#039; is the assembly accession number from NCBI, and the Biosample Accession is the Biosample accession number from NCBI.                                         &#039;&#039;&#039;a &amp;amp; b)&#039;&#039;&#039; &#039;&#039;&#039;HIVE IDs&#039;&#039;&#039; are accessions that are already uploaded into HIVE and the algorithm will automatically select this information as opposed to pulling information from outside sources. ]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Where to locate NCBI Information for the inputs:&#039;&#039; ==&lt;br /&gt;
Navigate to the NCBI legacy assembly page for your organism. Here you can find all of the information to be used for your computations.&lt;br /&gt;
[[File:Screenshot fig 3.png|none|thumb|518x518px|&#039;&#039;&#039;Figure 3.&#039;&#039;&#039; The information shown in the NCBI legacy assembly page. The RefSeq assembly accession corresponds to our organism of interest and will be used to fill out the &#039;&#039;&#039;“Assembly ID”&#039;&#039;&#039; input field on the HIVE3 ARGOS_QC input page. Please note that the bioproject matches the accession for the FDA_ARGOS bioproject, and there are 2 sequencing technologies listed, meaning that there will most likely be two SRR submissions that we can find on the SRA page (see Figure 5 and 6). ]]&lt;br /&gt;
[[File:Screenshot fig 4.png|none|thumb|670x670px|&#039;&#039;&#039;Figure 4&#039;&#039;&#039;. The bottom section of the legacy assembly page for our test organism. The column labeled RefSeq sequence lists the DNA RefSeq for our test organism which we will use for the “&#039;&#039;&#039;Reference Accession&#039;&#039;&#039;” field on the HIVE3 ARGOS_QC input page. ]]&lt;br /&gt;
[[File:Screenshot fig 5.png|none|thumb|625x625px|&#039;&#039;&#039;Figure 5.&#039;&#039;&#039; Clicking on the Biosample accession number seen in Figure 3 will redirect you to the NCBI Biosample page for this organism. Under &#039;&#039;&#039;“Related Information”&#039;&#039;&#039;, click &#039;&#039;&#039;“SRA”&#039;&#039;&#039; to navigate to the NCBI SRA page for this biosample.]]&lt;br /&gt;
[[File:Screenshot fig 6.png|none|thumb|634x634px|&#039;&#039;&#039;Figure 6.&#039;&#039;&#039; The SRA page lists different sequencing links. Each link is reported to be sequenced on different platforms; either for Illumina or for the PacBio platform. This is common. The methodology behind using the different platforms is to gather insight for the assembly at different perspectives and levels. Illumina sequences DNA as multiple short reads that can be used to create an accurate reconstruction of the genomic sections analyzed by estimating the average/best fit nucleotide sequence. PacBio is a long read sequencer that takes “movies” of the DNA sequence as it moves through the technology and captures the sequence in one go from start to finish. The long read sequence then acts as a map for the short accurate short read sequences that need to be assembled. Therefore, it is important to use all of the links reported in our QC pipeline.]]&lt;br /&gt;
[[File:Screenshot fig 7.png|none|thumb|635x635px|&#039;&#039;&#039;Figure 7.&#039;&#039;&#039; Clicking on the bottom link under the &#039;&#039;&#039;Runs&#039;&#039;&#039; section from the SRA page shown in &#039;&#039;&#039;Figure 6&#039;&#039;&#039; will redirect to the page containing the SRR file. Copy and paste the SRR accession number into the input field in the HIVE3 pipeline labeled “SRR”. You will need to do this for both (or more) SRR accessions. Check that the bioproject, biosample, and organism name all coincide with our test organism.]]&lt;br /&gt;
&lt;br /&gt;
= Single QC Computation =&lt;br /&gt;
A single QC computation will allow for assemblyQC, biosampleQC, and ngsQC to be performed on one organism with one assembly, but can include multiple SRR ids. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Input the name of the computation and the name of the folder you wish to store the computation in. If you would like to add to a pre-existing folder, input the exact folder name in the folder input field. It is case sensitive. See input criteria at the beginning of this protocol.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; Under the dropdown menu for Reads, select which input you will use. &lt;br /&gt;
&lt;br /&gt;
* For SRRs, type or paste in the SRR id. If there is more than one SRR id, click on the gray + sign to populate a new input field or use a , .&lt;br /&gt;
** Troubleshooting: if the computation fails, try removing the spaces in between the commas and SRR ids. No spaces.&lt;br /&gt;
* For HIVE IDs, click on the HIVE ID option from the dropdown menu. Click on the gray dropdown menu arrow next to HIVE reads. A pop-up window will open, as seen in Figure 9. &lt;br /&gt;
** Click on the ids that you wish to use in the computation. Use Ctrl + shift to highlight multiple ids.&lt;br /&gt;
&lt;br /&gt;
[[File:Fig8.png|none|thumb|700px|&#039;&#039;&#039;Figure 8.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline filled out for a single QC computation of our test organism, Salmonella enterica. ]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:Fig9.png|none|thumb|700px|&#039;&#039;&#039;Figure 9.&#039;&#039;&#039; Pop-up window for SRR HIVE ID selection. ]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Next to Reference, select the input object you would like to use for the computation.&lt;br /&gt;
&lt;br /&gt;
* For Reference Accession, type or paste in the id you wish to use.&lt;br /&gt;
* For Assembly IDs, type or paste in the id you wish to use. &lt;br /&gt;
* For HIVE Genomes, refer to Step 2 above on how to select a HIVE id. It is the same process.&lt;br /&gt;
* Refer to the beginning of this protocol for what ids can be inputted.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Next under BioSample Accessions, paste in the biosample ID you would like to use for the computation. This is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Next, select from the Coding Table dropdown menu the genetic code you would like to use for your computation, if applicable. The default is “human, viral (Standard)”.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Lastly, navigate to the Advanced tab and open the pipeline section. Under the All objects section, select argos-cflow System Application.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step7:&#039;&#039;&#039; Once all of the information has been inputted correctly, click on the big blue Submit button in the middle, as seen in Figure 8.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 8:&#039;&#039;&#039; Once your computation has been submitted, you can click on the Home tab found in the top left to go back to the homepage and see the workflow.&lt;br /&gt;
&lt;br /&gt;
= Batch Mode Computation =&lt;br /&gt;
Batch mode operates by a user-specified ratio of groups. With the help of semicolons and commas, the ratio would be 1:1:1 for a batch mode computation. It is 1:1:1, because we are clustering them by ; so the pipeline recognizes the ids between the ; as one computation. It would be one cluster of SRRs to one assembly to one biosample, that is one computation. There is a colorful and highlighted example below displaying the syntax for the inputs.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;They would be grouped for computations like this example:&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;u&amp;gt;Assembly 1:&amp;lt;/u&amp;gt; &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
&amp;lt;u&amp;gt;Assembly 2:&amp;lt;/u&amp;gt; &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
To separate between batches, use a semicolon “;” in between the IDs. A comma denotes separate IDs but semicolons as separate batches. These will be inputted in the General tab of the pipeline, same as single computation. Within each field, this is how the above example would look like in batch mode:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;u&amp;gt;SRR IDS:&amp;lt;/u&amp;gt; SRR0123456, SRR0123457, SRR0123458, SRR0123459&amp;lt;mark&amp;gt;;&amp;lt;/mark&amp;gt; SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
&amp;lt;u&amp;gt;Assembly IDs:&amp;lt;/u&amp;gt; GCA_0011223344.1&amp;lt;mark&amp;gt;;&amp;lt;/mark&amp;gt; GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
&amp;lt;u&amp;gt;BioSample Accessions:&amp;lt;/u&amp;gt; SAMN110654321, SAMN110654322&amp;lt;mark&amp;gt;;&amp;lt;/mark&amp;gt; SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Notice the semicolon separation according to the example above. The commas separate the ids, and the semicolon the batch.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Troubleshooting Note:&#039;&#039;&#039; If your computation fails or if there is an error, remove the spaces between the , and ; .Previously, this has thrown an error but has been fixed, but worth a shot if your computation fails. It would look like:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456,SRR0123457,SRR0123458,SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039;SRR0123451,SRR0123452,SRR0123453,SRR0123454&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Navigate to the tab title Batch, Figure 10. This can be found on the ARGOS input settings page, Figure 1., next to the General tab. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; For the parameter “batch service&amp;quot; at the bottom select, from the dropdown menu batch mode. This will have the pipeline set to Batch Mode rather than single computations.&lt;br /&gt;
&lt;br /&gt;
[[File:Fig.10.png|none|thumb|700px|&#039;&#039;&#039;Figure 10.&#039;&#039;&#039; Batch mode input settings window.]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Selecting the parameters. Click on the drop down menu next to the text “Parameter list”.&lt;br /&gt;
&lt;br /&gt;
* Use the black plus button next to ‘Parameter List’ to populate an entry field.&lt;br /&gt;
* Select from the dropdown field the correct parameter based on the input field you used in the general input page. This can be seen in Figure 11.&lt;br /&gt;
* For example, if you pasted in SRR Ids you would choose the parameter SRR IDs. If you chose HIVEIDs you would select HIVE IDs from the dropdown.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:Fig11.png|none|thumb|700px|&#039;&#039;&#039;Figure 11.&#039;&#039;&#039; Dropdown menu from the Batch tab displaying the parameter options. Select the batch parameter option based on what the input information is in the general tab. For example, if you entered SRR ids, select SRR IDs from the dropdown. If you input a reference ID, select Reference IDs.]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Input the ratio for the batch service. &lt;br /&gt;
&lt;br /&gt;
* For computations in batch mode in the one-click pipeline, the computations are separated by semicolon “;” and the IDs within the computations by a comma “,”. Since the workflow will parse the computations and recognize the IDs between the “;” as one computation, the ratio will be 1:1:1. &lt;br /&gt;
* If the ratio is 1:1:1 then enter the value 1 for each box.&lt;br /&gt;
* One set of SRRs to one assembly to one biosample. &lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Navigate to the Advanced tab and open the Pipeline section. As shown in Figure 12, click on the dropdown arrow next to Pipeline, and under All Projects, locate argos-cflow System Application. As shown in FIgure 13, select argos-cflow System Application by clicking on it, then confirm the selection by clicking Select at the bottom of the window. Leave all other settings in the Advanced section as default.&lt;br /&gt;
[[File:Fig12.png|none|thumb|700px|&#039;&#039;&#039;Figure 12.&#039;&#039;&#039; Advanced tab showing the Pipeline dropdown menu where the argos-cflow System Application can be selected.]]&lt;br /&gt;
[[File:Fig13.png|none|thumb|700px|&#039;&#039;&#039;Figure 13.&#039;&#039;&#039; All Projects window displaying available workflows. The argos-cflow System Application is highlighted and can be selected by clicking “Select” at the bottom. ]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Inputting the information correctly in each field. Navigate back to the input settings page, Figure 1. The same page that you had used for the single computations will be used, but the only difference will be the semicolons and commas. The example below will visually show you how the information will be inputted for a batch mode.&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Batch Mode Parameter breakdown:&#039;&#039; ==&lt;br /&gt;
[[File:Batcheg1.png|none|thumb|500px]] &lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;border:2px solid #e60000; border-collapse:collapse; margin:0 0 5px 0;&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:2px 6px;&amp;quot; | SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;border:2px solid #ff9900; border-collapse:collapse; margin:0 0 5px 0;&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:2px 6px;&amp;quot; | GCA_0011223344.1&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;border:2px solid #ffff00; border-collapse:collapse; margin:0 0 5px 0;&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:2px 6px;&amp;quot; | SAMN110654321, SAMN110654322&lt;br /&gt;
|}&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:Batcheg2.png|none|thumb|500px]]&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2:&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;border:2px solid #00ff00; border-collapse:collapse; margin:0 0 5px 0;&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:2px 6px;&amp;quot; | SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;border:2px solid #00ffff; border-collapse:collapse; margin:0 0 5px 0;&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:2px 6px;&amp;quot; | GCA_0011223345.1&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
{| style=&amp;quot;border:2px solid #8000ff; border-collapse:collapse; margin:0 0 5px 0;&amp;quot;&lt;br /&gt;
| style=&amp;quot;padding:2px 6px;&amp;quot; | SAMN110654323, SAMN110654324&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
Again, this is very similar to single computations, except that the batch mode will use semicolons and commas to separate the ids.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the input information is complete, hit the blue button Submit. You may exit the Argos pipeline window by hitting “Home” on the top left corner.&lt;br /&gt;
&lt;br /&gt;
= QC Computation Results =&lt;br /&gt;
Once you have submitted your computations, either single or batch, you should see the pipeline workflow in your Inbox or All Objects, as shown in Figure 12. You can also view the pipeline by clicking on the “workflows” tab also seen in Figure 14.&lt;br /&gt;
[[File:Fig14.png|none|thumb|700px|&#039;&#039;&#039;Figure 14.&#039;&#039;&#039;  The pipeline workflow displayed in the user’s inbox.]]&lt;br /&gt;
&lt;br /&gt;
As the workflow progresses, your computations will be stored in the folder that you named from the beginning of this protocol. To view the contents of the folder, simply click on the plus sign next to the folder or the folder name to open.&lt;br /&gt;
&lt;br /&gt;
Once your computations are complete the QC outputs will be stored in JSON file format from the computation “&#039;&#039;&#039;Post-Alignment Quality Controls&#039;&#039;&#039;” or under the &#039;&#039;&#039;“CFlow”&#039;&#039;&#039; workflow. P-A QC can be found in the folder you specified for the computation or CFlow in All Objects. To view the JSONs click on the name so that it is highlighted blue and click on the tab from the bottom menu named “Available Downloads”.&lt;br /&gt;
[[File:Fig15.png|none|thumb|700px|&#039;&#039;&#039;Figure 15.&#039;&#039;&#039; The available downloads tab and the 5 JSON files that are the QC outputs.]]&lt;br /&gt;
There will be 5 files reported in JSON format. Click the blue/green download icon next to each file to see the results. The file labeled &#039;&#039;&#039;qcAll.json&#039;&#039;&#039; will have our assemblyQC results. &#039;&#039;&#039;qcNGS.json&#039;&#039;&#039; will have our ngsQC results and &#039;&#039;&#039;biosample.json&#039;&#039;&#039; the biosample information. We currently do not submit qcPos.json or refAnnot.json to the ARGOS DB, but the information is there to better help you understand your computation.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Projects&amp;diff=998</id>
		<title>Projects</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Projects&amp;diff=998"/>
		<updated>2025-09-08T19:21:23Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{DISPLAYTITLE:&amp;lt;span style=&amp;quot;position: absolute; clip: rect(1px 1px 1px 1px); clip: rect(1px, 1px, 1px, 1px);&amp;quot;&amp;gt;{{FULLPAGENAME}}&amp;lt;/span&amp;gt;}}&lt;br /&gt;
__NOTOC__&lt;br /&gt;
&amp;lt;!-- BANNER ACROSS TOP OF PAGE --&amp;gt;&lt;br /&gt;
&amp;lt;div id=&amp;quot;ggw-topbanner&amp;quot; style=&amp;quot;clear:both; position:relative; box-sizing:border-box; width:100%; margin:1.2em 0 6px; min-width:47em; border:1px solid #ddd; background-color:#f9f9f9; color:#000;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;margin:0.4em; text-align:center;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;font-size:160%; padding:.1em;&amp;quot;&amp;gt;Current Projects&amp;lt;/div&amp;gt;&lt;br /&gt;
        &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;clear: both;&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;div id=&amp;quot;ggw_row2&amp;quot; style=&amp;quot;display: flex; flex-flow: row wrap; justify-content: space-between; padding: 0; margin: 0 -5px 0 -5px;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://hive.biochemistry.gwu.edu/dna.cgi?cmd=main The High-performance Integrated Virtual Environment (HIVE) platform]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
HIVE is a cloud-based environment optimized for the storage and analysis of extra-large data, such as biomedical data, clinical data, next-generation sequencing (NGS) data, mass spectrometry files, confocal microscopy images, post-market surveillance data, medical recall data, and many others. HIVE provides secure web access for authorized users to deposit, retrieve, annotate and compute on Big Data, and analyze the outcomes using web user interfaces. [https://docs.google.com/document/d/1F5iq00uKkJfdSsbwanvKOy-nPnwijH56mwbwa_HhzfY/edit?tab=t.0#heading=h.7dlfmngwfzih More here].&lt;br /&gt;
&lt;br /&gt;
The HIVE platform and associated algorithms such as CensuScope and HIVE-Hexagon is used to support Metgenomics analysis infrastructure.&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
[[GW-HIVE WIKI]]&lt;br /&gt;
&lt;br /&gt;
[[METAGENOMICS WIKI]]&lt;br /&gt;
        &lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
	&amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://data.argosdb.org/ FDA-ARGOS Project (Food and Drug Administration-dAtabase for Regulatory-Grade micrObial Sequences)]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
The FDA-ARGOS Project (Food and Drug Administration-dAtabase for Regulatory-Grade micrObial Sequences) is a collaborative effort to create a high-quality genomic database for identifying and characterizing microbial pathogens. Developed in partnership with the FDA, University of Maryland, and NCBI, the project provides regulatory-grade genomic data, crucial for public health and diagnostic use. Expanded in 2021 with support from GWU, Temple University, and Embleema, FDA-ARGOS aims to enhance infectious disease research through rigorous quality control protocols. The ArgosDB hosts this data, offering downloadable sequences and reproducible workflows for research and regulatory applications.[https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos More here].&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
[[FDA-ARGOS WIKI]]&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://www.biocomputeobject.org/ BioCompute Objects (BCO)]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
The BioCompute is FDA funded project to establish a framework for community-based development of standards for harmonization of High-throughput Sequencing (HTS), standardization of data formats, promotion of interoperability, and bioinformatics verification protocols. The BioCompute Object (BCO) was developed in the High-throughput Sequencing Computational Standards for Regulatory Sciences (HTS-CSRS) initiative in the BioCompute Objects Portal (BOP), a web portal to serve as a collaborative ground to encourage a dialogue to facilitate interoperability between different bioinformatic pipelines, industries, and developers. HIVE capabilities have been leveraged to support the development of the BCO. The BCO is versatile and adaptable to other common HTS analysis platforms. [https://docs.google.com/document/d/1WQFZm_PFiQXob4NyOKq6y-2ywnbmNoFHSS27fYf3l4Y/edit?tab=t.0#heading=h.bs8eki17tykx More here].&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
[https://wiki.biocomputeobject.org/Main_Page BIOCOMPUTE OBJECTS WIKI]&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://www.glygen.org/ GlyGen]&amp;lt;/h3&amp;gt;&lt;br /&gt;
	&amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
GlyGen (gly-glycobiology; gen-information), [https://www.glygen.org/&amp;lt;nowiki&amp;gt;] is an advanced glycoinformatics resource developed to facilitate discovery in basic and translational glycobiology research along with enhancing the integration of multidisciplinary information from diverse resources. GlyGen includes knowledge about molecular, biophysical and functional properties of glycans, genes, and proteins organized in pathways and ontologies, plus a rapidly growing body of biological big data related to cancer mutation and expression. GlyGen adopts an innovative user-driven approach for implementing, prioritizing and knowledge disseminating tools to address the questions and needs of glycobiology community. GlyGen is funded by the National Institute of General Medical Sciences under the grant # 1R24GM146616 - 01 and the  National Institutes of Health Office of Strategic Coordination - The Common Fund under the grant # 1OT2OD032092. More information about GlyGen - &amp;lt;/nowiki&amp;gt;https://www.glygen.org/about/ &amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
[https://wiki.glygen.org/Main_Page GlyGen WIKI]&lt;br /&gt;
        &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&amp;lt;div id=&amp;quot;ggw_row3&amp;quot; style=&amp;quot;display: flex; flex-flow: row wrap; justify-content: space-between; padding: 0; margin: 0 -5px 0 -5px;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://hivelab.biochemistry.gwu.edu/predictmod PredictMod]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
PredictMod is an application designed to predict the outcome of an intervention prior to a patient initiating treatment. Our goal is to provide clinicians with a powerful decision making tool that enhances clinical understanding of patient-level data. The PredictMod platform utilizes machine learning tools and complex datasets based on electronic health records, gut microbiome, and -omics data to forecast patient outcomes, often in response to treatment for a particular condition. While our primary condition of interest is Prediabetes, the tool is designed to be used for a variety of conditions, interventions, and data types.  &amp;lt;br&amp;gt; &amp;lt;br&amp;gt;&lt;br /&gt;
[https://hivelab.biochemistry.gwu.edu/wiki/PredictMod PredictMod WIKI]&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[[GW-FEAST]]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
The GW Federated Ecosystems for Analytics and Standardized Technologies (GW-FEAST) project is part of the ARPA-H FEAST performer team initiative that includes academic and industry partners. The goal of the ARPA-H performer teams is “to create bridges across data silos to make health data more accessible and usable”. &amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
[https://hivelab.biochemistry.gwu.edu/wiki/GW-FEAST GW-FEAST WIKI]&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://hivelab.biochemistry.gwu.edu/biomarker-partnership Biomarker Knowledgebase]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
The Biomarker Partnership is a CFDE sponsored project to develop a knowledgebase that will organize and integrate biomarker data from different public sources. The data will be connected to contextual information to show a novel systems-level view of biomarkers. The motivation for this project is to improve the harmonization and organization of biomarker data. This will be done by mapping biomarkers from public sources to, and across, CF data elements. This mapping will bridge knowledge across multiple DCCs and biomedical disciplines.&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
[https://wiki.biomarkerkb.org/Main_Page BioMarkerKB WIKI]&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
{{DISPLAYTITLE:&amp;lt;span style=&amp;quot;position: absolute; clip: rect(1px 1px 1px 1px); clip: rect(1px, 1px, 1px, 1px);&amp;quot;&amp;gt;{{FULLPAGENAME}}&amp;lt;/span&amp;gt;}}&lt;br /&gt;
__NOTOC__&lt;br /&gt;
&amp;lt;!-- BANNER ACROSS TOP OF PAGE --&amp;gt;&lt;br /&gt;
&amp;lt;div id=&amp;quot;ggw-topbanner&amp;quot; style=&amp;quot;clear:both; position:relative; box-sizing:border-box; width:100%; margin:1.2em 0 6px; min-width:47em; border:1px solid #ddd; background-color:#f9f9f9; color:#000;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;margin:0.4em; text-align:center;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;font-size:160%; padding:.1em;&amp;quot;&amp;gt;Past Projects&amp;lt;/div&amp;gt;&lt;br /&gt;
        &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;clear: both;&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;div id=&amp;quot;ggw_row2&amp;quot; style=&amp;quot;display: flex; flex-flow: row wrap; justify-content: space-between; padding: 0; margin: 0 -5px 0 -5px;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://hivelab.tst.biochemistry.gwu.edu/gfkb Gut Microbiome Analytic System (Microbiome)]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
The HIVE team received NSF funding to develop a Gut Microbiome Monitoring System (GutFeeling) as a tool which when used over time will allow users to rectify their dietary (such as consumption of probiotics and prebiotics) and other lifestyle habits and to help restore their normal microbiome. Rapid analysis of the large amount of metagenomic data, a major bottleneck, has been resolved by our group through the development of a novel algorithm and accompanying software called CensuScope. Through analysis of healthy gut microbiome data, we are actively developing a Knowledge Base (GutFeelingKB) to provide a clearer picture of not only an ideal personalized microbiome but also establish baseline characteristics for each customer. The Mazumder Lab is collaborating with the Milken School of Public Health and Kamtek Sequencing Facility to investigate the relationship between bacterial species commonly present in the digestive tract, diet, physical activity, lifestyle habits, and metabolic risk factors. [https://docs.google.com/document/d/18WyVTJrrf-FR0sHt634vO8Lwel-4OQxP9sNar7gYYro/edit?tab=t.0#heading=h.7qbm3f7lky31 More here].&lt;br /&gt;
        &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
	&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;HIVE-EQAPOL Project on HIVE NGS Data Processing and Analysis&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
For this project, our group works closely with the External Quality Assurance Program Oversight Laboratory (EQAPOL) team to conduct HIV NGS data analysis and collaborate in terms of analyzing, storing, and tracking HIV NGS Data. Reliable identification of strains is critical for developing new assays, validating assay platforms, assisting regulators to evaluate test kits, monitoring HIV drug resistance, and informing vaccine development. The HIVE tools and platform are used for virus identification, recombination analysis, and clone discovery.&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://www.oncomx.org/ OncoMX]&amp;lt;/h3&amp;gt;&lt;br /&gt;
	&amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
The OncoMX mission is to create an integrated cancer mutation and expression resource for exploring cancer biomarkers. OncoMX is a collaboration between the George Washington University (GW), NASA&#039;s Jet Propulsion Laboratory (JPL), the Swiss Institute of Bioinformatics (SIB), and the University of Delaware (UD). The core knowledgebase of OncoMX is derived from BioMuta and BioXpress integrated cancer mutation and expression databases which are actively maintained. Normal expression data from Bgee and custom text mining software augment the cancer data to improve functional interpretation of the reported variants and expression profiles. All data are wrapped into the OncoMX database and web portal, mapped to additional functional information from NCI Early Detection Research Network (EDRN) and Reactome. It is expected that the large-scale integration of cancer data and supporting information, provided by OncoMX with direct community feedback, will benefit cancer research by improving synthesis of information and may make earlier detection a reality.&lt;br /&gt;
        &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[https://hive.biochemistry.gwu.edu/dna.cgi?cmd=main Glycoproteomics Characterization Workflow and Data-Analysis Pipeline for Vaccines and Biosimilars]&amp;lt;/h3&amp;gt;&lt;br /&gt;
	&amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
In this FDA funded project we are extending High-performance Integrated Virtual Environment (HIVE) capabilities through the development and integration of software tools and datasets for comparative analysis of glycoproteins. Glycomic analysis has many angles and has been extensively reviewed in recent literature. We propose to rely on the independent development of the glycomics field and incorporate these approaches in the HIVE pipeline as they mature while we develop a standardized glycoinformatics pipeline that will benefit investigators and regulators at the FDA.&lt;br /&gt;
        &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
{{DISPLAYTITLE:&amp;lt;span style=&amp;quot;position: absolute; clip: rect(1px 1px 1px 1px); clip: rect(1px, 1px, 1px, 1px);&amp;quot;&amp;gt;{{FULLPAGENAME}}&amp;lt;/span&amp;gt;}}&lt;br /&gt;
__NOTOC__&lt;br /&gt;
&amp;lt;!-- BANNER ACROSS TOP OF PAGE --&amp;gt;&lt;br /&gt;
&amp;lt;div id=&amp;quot;ggw-topbanner&amp;quot; style=&amp;quot;clear:both; position:relative; box-sizing:border-box; width:100%; margin:1.2em 0 6px; min-width:47em; border:1px solid #ddd; background-color:#f9f9f9; color:#000;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;margin:0.4em; text-align:center;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;font-size:160%; padding:.1em;&amp;quot;&amp;gt;RESOURCES&amp;lt;/div&amp;gt;&lt;br /&gt;
        &amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;div style=&amp;quot;clear: both;&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;div style=&amp;quot;clear: both;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&amp;lt;div id=&amp;quot;ggw_row3&amp;quot; style=&amp;quot;display: flex; flex-flow: row wrap; justify-content: space-between; padding: 0; margin: 0 -5px 0 -5px;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[[Tool Resources]]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
&amp;amp;nbsp;&amp;amp;nbsp;&amp;amp;nbsp;&amp;amp;nbsp;&#039;&#039;&#039;&#039;&#039;Main article:&#039;&#039;&#039; [[Tool Resources]]&#039;&#039;&amp;lt;br&amp;gt;There are a variety of bioinformatic tool resources developed by our team.&lt;br /&gt;
&amp;lt;br /&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;div style=&amp;quot;clear: both;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&amp;lt;div id=&amp;quot;ggw_row3&amp;quot; style=&amp;quot;display: flex; flex-flow: row wrap; justify-content: space-between; padding: 0; margin: 0 -5px 0 -5px;&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;div style=&amp;quot;flex: 1; margin: 5px; min-width: 210px; border: 1px solid #CCC;	padding: 0 10px 10px 10px; box-shadow: 0 2px 2px rgba(0,0,0,0.1); background: #f5faff;&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;h3&amp;gt;[[Dataset Resources]]&amp;lt;/h3&amp;gt;&lt;br /&gt;
        &amp;lt;div style=&amp;quot;border-top: 1px solid #CCC; padding-top: 0.5em;&amp;quot;&amp;gt;&lt;br /&gt;
&amp;amp;nbsp;&amp;amp;nbsp;&amp;amp;nbsp;&amp;amp;nbsp;&#039;&#039;&#039;&#039;&#039;Main article:&#039;&#039;&#039; [[Dataset Resources]]&#039;&#039;&amp;lt;br&amp;gt;There are a variety of bioinformatic dataset resources integrated by our team.&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
    &amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=991</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=991"/>
		<updated>2025-09-08T17:19:14Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Project Publications */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
* FDA-ARGOS NCBI BioProject can be found [https://www.ncbi.nlm.nih.gov/bioproject/231221 here].&lt;br /&gt;
* [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here]. This is a static webpage that will be updated with each push.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&lt;br /&gt;
== ARGOS Usage Tutorials and Protocols ==&lt;br /&gt;
&lt;br /&gt;
The ARGOS One-Click Pipeline is used to create the QC metrics and results displayed in our data tables on data.argosdb.org. The ARGOS One-Click Pipeline Usage tutorial can be found [[ARGOSQC Usage Tutorial|here]]. &lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&lt;br /&gt;
*Coming Soon: Journal Publication highlighting our QC metrics and data dictionary for the ARGOS Database&amp;lt;br /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=990</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=990"/>
		<updated>2025-09-08T17:15:58Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* FDA-ARGOS Database */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
* FDA-ARGOS NCBI BioProject can be found [https://www.ncbi.nlm.nih.gov/bioproject/231221 here].&lt;br /&gt;
* [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here]. This is a static webpage that will be updated with each push.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&lt;br /&gt;
== ARGOS Usage Tutorials and Protocols ==&lt;br /&gt;
&lt;br /&gt;
The ARGOS One-Click Pipeline is used to create the QC metrics and results displayed in our data tables on data.argosdb.org. The ARGOS One-Click Pipeline Usage tutorial can be found [[ARGOSQC Usage Tutorial|here]]. &lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&amp;lt;br /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=989</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=989"/>
		<updated>2025-09-08T17:15:14Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* ARGOS Usage Tutorials and Protocols */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
* FDA-ARGOS NCBI BioProject can be found [https://www.ncbi.nlm.nih.gov/bioproject/231221 here].&lt;br /&gt;
* [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here].&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&lt;br /&gt;
== ARGOS Usage Tutorials and Protocols ==&lt;br /&gt;
&lt;br /&gt;
The ARGOS One-Click Pipeline is used to create the QC metrics and results displayed in our data tables on data.argosdb.org. The ARGOS One-Click Pipeline Usage tutorial can be found [[ARGOSQC Usage Tutorial|here]]. &lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&amp;lt;br /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=988</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=988"/>
		<updated>2025-09-08T17:13:20Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
* FDA-ARGOS NCBI BioProject can be found [https://www.ncbi.nlm.nih.gov/bioproject/231221 here].&lt;br /&gt;
* [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here].&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&lt;br /&gt;
== ARGOS Usage Tutorials and Protocols ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&amp;lt;br /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=987</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=987"/>
		<updated>2025-09-08T17:12:38Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
* FDA-ARGOS NCBI BioProject can be found [https://www.ncbi.nlm.nih.gov/bioproject/231221 here].&lt;br /&gt;
* [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here].&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;br /&gt;
&lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&amp;lt;br /&amp;gt;&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOS_Contact_Us&amp;diff=986</id>
		<title>ARGOS Contact Us</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOS_Contact_Us&amp;diff=986"/>
		<updated>2025-09-08T17:11:49Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;To contact us for any inquires, &lt;br /&gt;
* email mazumder_lab@gwu.edu&lt;br /&gt;
Feel free to reach out to us about the data published in data.argosdb.org or any questions you may have.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
To return to the ARGOS Main Page, [[FDA-ARGOS WIKI|click here]].&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=985</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=985"/>
		<updated>2025-09-08T17:10:44Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Introduction */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information from FDA Announcements]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
* FDA-ARGOS NCBI BioProject can be found [https://www.ncbi.nlm.nih.gov/bioproject/231221 here].&lt;br /&gt;
* [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here].&lt;br /&gt;
&lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_2025&amp;diff=976</id>
		<title>Volunteership 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_2025&amp;diff=976"/>
		<updated>2025-08-29T14:12:03Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Volunteers */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;For Fall opportunities, [[Volunteership Fall 2025|click here to view our Fall 2025 Volunteership Program]].&amp;lt;h2&amp;gt;2025 Volunteer Program Details&amp;lt;/h2&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h3&amp;gt;Dates&amp;lt;/h3&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;Volunteer Zoom Kick-Off Meeting&amp;lt;/strong&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
May 27, 2025 | 3:30 to 4:30 PM&lt;br /&gt;
&lt;br /&gt;
&amp;lt;strong&amp;gt;Program Dates: June 2nd, 2025 – July 25th, 2025&amp;lt;/strong&amp;gt; (8 weeks)&amp;lt;br&amp;gt;&lt;br /&gt;
Monday to Friday | Remote | No breaks&lt;br /&gt;
&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h3&amp;gt;Volunteer Expectations&amp;lt;/h3&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;Daily progress updates via Slack (scrum).&amp;lt;/li&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;Regular Zoom meetings with the assigned project point of contact.&amp;lt;/li&amp;gt;&amp;lt;li&amp;gt;Expected to dedicate 5–6 hours per day to project work, with the remaining time focused on skill development or reading. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;p style=&amp;quot;color: red;&amp;quot;&amp;gt;&amp;lt;strong&amp;gt;Important:&amp;lt;/strong&amp;gt; If the scrum is not updated for 2 consecutive days, the candidate will be &amp;lt;u&amp;gt;automatically dropped&amp;lt;/u&amp;gt; from the program.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h3&amp;gt;Potential Projects&amp;lt;/h3&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;BiomarkerKB ([https://biomarkerkb.org biomarkerkb.org]) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&amp;lt;/li&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;GlyGen ([https://glygen.org glygen.org]) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information. &amp;lt;/li&amp;gt;&amp;lt;li&amp;gt;ARGOS ([https://argosdb.org argosdb.org]) project: Analyze genomics data using HIVE to identify reference genome assemblies. &amp;lt;/li&amp;gt;&amp;lt;li&amp;gt;PredictMod ([https://hivelab.biochemistry.gwu.edu/predictmod hivelab.biochemistry.gwu.edu/predictmod]) project. Identifying datasets and harmonizing them so that they can be used to generate ML models.  &amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h4&amp;gt;1. BiomarkerKB Biocuration Project Ideas&amp;lt;/h4&amp;gt;POC: Daniall Masood, Maria Kim&lt;br /&gt;
# Curate biomarkers for a specific disease (Alzheimers)&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer&lt;br /&gt;
&lt;br /&gt;
Data Identification &amp;amp; Curation: &lt;br /&gt;
&lt;br /&gt;
# Identify publicly-available datasets from scientific literature that can be used for intervention outcome prediction models.&lt;br /&gt;
# Curate indicators of useful ML publications that could be used to train an LLM to recommend relevant publications for cancer modeling.&lt;br /&gt;
&lt;br /&gt;
Modeling &amp;amp; Integration (for those with experience in programming/ML)&lt;br /&gt;
&lt;br /&gt;
# Conduct data harmonization and pre-processing following established project pipelines to make ML-ready dataset and data dictionary.&lt;br /&gt;
# Perform model training and document ML pipeline in a BioCompute Object (BCO).&lt;br /&gt;
# Integrate model into PredictMod platform.&lt;br /&gt;
&lt;br /&gt;
Individuals with a background or interest in machine learning should reach out to lorikrammer@gwu.edu with a potential dataset to determine if it is a feasible project for the summer.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
# Update data tables for more efficient computations&lt;br /&gt;
## Student would review and input additional data and IDs in the tables/sheets used to perform computations. This would be manual work (but super important), but would require high attention to detail. ~1 week&#039;s worth of work&lt;br /&gt;
## Requires Python/shell coding background. Student would run scripts that prepare and format data tables that are pushed to data.argosdb.org. Coding knowledge is needed in case of errors, bugs, or other mishaps in the code. Ongoing work as computations are performed.&lt;br /&gt;
# Curate and report on current pathogens to upload to ARGOS&lt;br /&gt;
## Student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found. ~4-10 weeks worth of work&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&amp;lt;hr&amp;gt;&lt;br /&gt;
&amp;lt;h3&amp;gt;Requirements for Completion&amp;lt;/h3&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;strong&amp;gt;Note:&amp;lt;/strong&amp;gt; The following are &amp;lt;u&amp;gt;mandatory&amp;lt;/u&amp;gt;. Failure to complete any will result in an incomplete volunteer record.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h4&amp;gt;Documentation&amp;lt;/h4&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h4&amp;gt;Written Report&amp;lt;/h4&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h4&amp;gt;Presentation &amp;amp; Slide Submission&amp;lt;/h4&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Present your work last week of the 8-week period.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Slides must be submitted to the Admin Team and should include:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;See Symposium Slides Guidelines below&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
Contact the Admin Team to access previously submitted slides.&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
=== Volunteers ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
|-&lt;br /&gt;
! Name&lt;br /&gt;
!Project&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
| [https://www.linkedin.com/in/gracesjchong/ Grace Chong]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
# PredictMod&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/alma-ogunsina-4959072b1/ Alma Ogunsina]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
# ARGOS&lt;br /&gt;
# PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/diya-kamalabharathy-62557935a/ Diya Kamalabharathy]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# PredictMod Machine Learning&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/harivinay-prasad-reddy-gujjula-a06ba71bb/ Harivinay P. Gujjula]&lt;br /&gt;
|GlyGen curation&lt;br /&gt;
|&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# BioMarkerKB Biocuration&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;Arhamur Rauf&#039;&#039;&#039;&lt;br /&gt;
|ARGOS&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/miao-wang-88b602290/Miao&amp;amp;#x20;Wang Miao Wang]&lt;br /&gt;
|ARGOS&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration Project Ideas&lt;br /&gt;
# FDA-ARGOS Computation and Pathogen Curation Project&lt;br /&gt;
# PredictMod Machine Learning Project Ideas&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/nahom-gebreselassie-1545ab336/ Nahom Abel]&lt;br /&gt;
|GlyGen curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/kajal-patel-cs/ Kajal Sanjaykumar Patel]&lt;br /&gt;
|GlyGen and PubMed project&lt;br /&gt;
|&lt;br /&gt;
#PredictMod&lt;br /&gt;
#BiomarkerKB&lt;br /&gt;
#GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/john-mccaffrey-b8850930a/ John McCaffrey]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# PredictMod&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
# GlyGen Biocuration &lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/nathan-ressom/ Nathan Ressom]&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
# PredictMod &lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/aaron-ressom/ Aaron Ressom] &lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB &lt;br /&gt;
# PredictMod &lt;br /&gt;
# GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/akale-kinfe/ Akale Kinfe]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/aise-arpinar-a8bb9b373/?original_referer= Aise Arpinar]&lt;br /&gt;
|GlyGen curation&lt;br /&gt;
|&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# GlyGen Publication Analysis&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/piyush-pandey-906b582b5/ Piyush Pandey]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration &lt;br /&gt;
# PredictMod &lt;br /&gt;
# GlyGen Biocuration &lt;br /&gt;
|-&lt;br /&gt;
|[http://www.linkedin.com/in/filmawit-zeru-203272363 Filmawit Zeru]&lt;br /&gt;
|GlycoSiteMiner project&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
# GlyGen&lt;br /&gt;
# ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/mathias-belay-03b51a2a3/ Mathias Belay]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# GlyGen&lt;br /&gt;
# PredictMod&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/isaac-kim-b644bb231/ Isaac Kim]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
# PredictMod&lt;br /&gt;
# GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/sohana-bahl-6549a2376/ Sohana Bahl]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/ana-vohralikova-794a4433a?utm_source=share&amp;amp;utm_campaign=share_via&amp;amp;utm_content=profile&amp;amp;utm_medium=ios_app Ana Vohralikova]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration Project&lt;br /&gt;
# GlyGen Biocuration Project&lt;br /&gt;
# FDA-ARGOS Computation and Pathogen&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Symposium Slide Guidelines ===&lt;br /&gt;
&#039;&#039;&#039;Content Clarity&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
   •       &#039;&#039;&#039;Keep It Simple:&#039;&#039;&#039; Use concise bullet points instead of long paragraphs. Aim for no more than 6 bullet points per slide. &lt;br /&gt;
&lt;br /&gt;
  •        &#039;&#039;&#039;Focus on Key Points:&#039;&#039;&#039; Highlight the main ideas or data you want your audience to remember. &lt;br /&gt;
&lt;br /&gt;
  •        &#039;&#039;&#039;Consistent Layout:&#039;&#039;&#039; Use a consistent layout for each slide, including fonts, colors, and background. This helps maintain a professional look. &lt;br /&gt;
&lt;br /&gt;
  •        &#039;&#039;&#039;High-Quality Images:&#039;&#039;&#039; Use high-resolution images and graphics to illustrate your points. Avoid using clip art. &lt;br /&gt;
&lt;br /&gt;
  •        &#039;&#039;&#039;Readable Fonts:&#039;&#039;&#039; Use easy-to-read fonts (e.g., Arial, Calibri) and ensure font sizes are large enough to be seen from a distance (24 pt or larger for main text). &lt;br /&gt;
&lt;br /&gt;
  •        &#039;&#039;&#039;Contrast:&#039;&#039;&#039; Ensure there is high contrast between text and background (e.g., dark text on a light background). &lt;br /&gt;
&lt;br /&gt;
  •        &#039;&#039;&#039;Citation:&#039;&#039;&#039; Cite a publication to support the information presented in proper citation format. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outline for Symposium presentation&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
1.       Introduction: &lt;br /&gt;
&lt;br /&gt;
2.       Project Descriptions &lt;br /&gt;
&lt;br /&gt;
3.       Objectives and Goals: &lt;br /&gt;
&lt;br /&gt;
4.       Methods, Results, Achievements and Contributions: &lt;br /&gt;
&lt;br /&gt;
5.       Future Plans: &lt;br /&gt;
&lt;br /&gt;
6.       Skills and Knowledge Gained: &lt;br /&gt;
&lt;br /&gt;
7.       Acknowledgments: &lt;br /&gt;
&lt;br /&gt;
8.       Q&amp;amp;A Session: &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Outline&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;1. Introduction:&#039;&#039;&#039;  (1 slide)&lt;br /&gt;
&lt;br /&gt;
  - Briefly introduce yourself.  &lt;br /&gt;
&lt;br /&gt;
  - Add your picture and name on the introduction slide.  If it is group add the group picture.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;2. Project Descriptions:&#039;&#039;&#039;  (1 slide)&lt;br /&gt;
&lt;br /&gt;
  - Provide context and background information about the project.  &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. Project Objectives and Goals:&#039;&#039;&#039;  (1 slide)&lt;br /&gt;
&lt;br /&gt;
  - Describe the main objectives of the project or initiative.  &lt;br /&gt;
&lt;br /&gt;
  - Discuss any additional goals or desired outcomes.  &lt;br /&gt;
&lt;br /&gt;
  - Explain why these objectives and goals are important.  &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;4. Methods, Results, Achievements and Contributions:&#039;&#039;&#039;  &lt;br /&gt;
&lt;br /&gt;
  -  Highlight the methods/tools used in the project.  &lt;br /&gt;
&lt;br /&gt;
  - Highlight the key results and outcomes of the project.  &lt;br /&gt;
&lt;br /&gt;
  - Discuss the most significant achievements and milestones reached.  &lt;br /&gt;
&lt;br /&gt;
  - Explain how each member of the team project contributed to the project (for group project) &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039; &#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. Future Plans&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
  - Next steps or future plans for the project&lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;6. Skills and Knowledge Gained:&#039;&#039;&#039;  (1 slide)&lt;br /&gt;
&lt;br /&gt;
  -   Detail any technical skills acquired or improved.  &lt;br /&gt;
&lt;br /&gt;
  - Highlight any soft skills, such as communication or teamwork, that were developed.  &lt;br /&gt;
&lt;br /&gt;
  - Discuss new knowledge gained in specific areas or subjects.  &lt;br /&gt;
&lt;br /&gt;
  -  Share any personal reflections on the experience and what was learned.  &lt;br /&gt;
&lt;br /&gt;
  - Discuss any challenges or obstacles encountered and how they were overcome.  &lt;br /&gt;
&lt;br /&gt;
  - Provide key insights or lessons learned from the project.  &lt;br /&gt;
&lt;br /&gt;
  &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;7. Acknowledgments:&#039;&#039;&#039;  &#039;&#039;&#039;:&#039;&#039;&#039;  (1 slide)&lt;br /&gt;
&lt;br /&gt;
  - Acknowledge the contributions of team members and collaborators.  &lt;br /&gt;
&lt;br /&gt;
- Recognize the guidance and support of mentors and advisors.  &lt;br /&gt;
&lt;br /&gt;
  - Acknowledge the Project Funding.  Eg. CFDE&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;8. Q&amp;amp;A Session:&#039;&#039;&#039;  &lt;br /&gt;
&lt;br /&gt;
  - Invite the audience to ask questions and engage in discussion.  &lt;br /&gt;
&lt;br /&gt;
  - Provide clear and thoughtful responses to audience questions.  &lt;br /&gt;
&lt;br /&gt;
  - Offer closing remarks and thank the audience for their participation. &lt;br /&gt;
&lt;br /&gt;
Note – If you have limited presentation time you can also merge few topics into one.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=938</id>
		<title>Volunteership Fall 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=938"/>
		<updated>2025-08-20T18:18:22Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* 4. PredictMod Machine Learning Project Ideas */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== 2025 Volunteer Program Details ==&lt;br /&gt;
&lt;br /&gt;
=== Dates ===&lt;br /&gt;
&#039;&#039;&#039;Application Deadline&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 22, 2025, Noon (email your updated resume and projects in order of preference). Acceptance letter/email will be sent to candidates latest the day after the kick-off meeting.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Volunteer Zoom Kick-Off Meeting&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 25, 2025 | 4:00 to 5:00 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Program Dates: September 1st, 2025 – November 30th, 2025&#039;&#039;&#039; (13 weeks)&lt;br /&gt;
&lt;br /&gt;
Remote | Hybrid for GW employees and students (Ross Hall 5th floor)&lt;br /&gt;
&lt;br /&gt;
[https://hivelab.biochemistry.gwu.edu/wiki/Volunteership_2025 Summer 2025 Volunteership] (Closed)&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteer Expectations ===&lt;br /&gt;
&lt;br /&gt;
# Minimum commitment of 10 hours per week.&lt;br /&gt;
# Progress updates via Slack at least 3 days per week (scrum).&lt;br /&gt;
# Regular Zoom meetings with the assigned project point of contact.&lt;br /&gt;
# Attend some lectures or seminars remotely (max 4-5).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;Important:&#039;&#039;&#039; If the scrum is not updated for 2 consecutive working days, the candidate will be automatically dropped from the program.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Potential Projects ===&lt;br /&gt;
We are excited to continue our bioinformatics volunteership program in Fall 2025. This program offers students the opportunity to work on bioinformatics projects supported by agencies such as the NIH, ARPA-H, and FDA. Participants will gain exposure to a variety of activities within a bioinformatics lab, including data analysis, computational biology, and genomics. If you are interested, please email mazumder_lab@gwu.edu your resume and a ranked list of the projects that interest you most. You can also indicate if you want to focus on specific areas that are of interest to you.&lt;br /&gt;
# BiomarkerKB (biomarkerkb.org) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&lt;br /&gt;
# GlyGen (glygen.org) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information.&lt;br /&gt;
# ARGOS (argosdb.org) project: Analyze genomics data using HIVE to identify reference genome assemblies.&lt;br /&gt;
# PredictMod (hivelab.biochemistry.gwu.edu/predictmod) project. Curating PMIDs for intervention outcome prediction dataset LLM recommendation training.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
==== 1. BiomarkerKB Biocuration Project Ideas ====&lt;br /&gt;
POC: Daniall Masood, Maria Kim&lt;br /&gt;
&lt;br /&gt;
# Curate biomarkers for a specific disease&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
# Continue working on LLM methods started by volunteers over the summer.&lt;br /&gt;
## The data is available as well as some preliminary research and work done by previous volunteers in this area.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer, Tianyi Wang, Pat McNeely (optional)&lt;br /&gt;
&lt;br /&gt;
Identifying relevant and useful publicly-available datasets for machine learning is currently a resource-intensive task. This curation project aims to develop a corpus for training an AI model to recommend PMIDs with publicly-available datasets useful for intervention outcome prediction models. The corpus will include an annotation spreadsheet + annotated PDFs for PubMed articles relevant to prostate, lung, breast, and liver cancer, and focus on indicators such as condition, intervention, and response.&lt;br /&gt;
&lt;br /&gt;
PMID curation involves:&lt;br /&gt;
&lt;br /&gt;
# Identify potentially relevant PMIDs that may have publicly-available datasets for training intervention outcome prediction models.&lt;br /&gt;
# Curate indicators of useful ML publications that could be used to train an LLM to recommend relevant publications for cancer modeling.&lt;br /&gt;
# Review peer curations and resolve annotation conflicts.&lt;br /&gt;
&lt;br /&gt;
Interested individuals should reach out to lorikrammer@gwu.edu.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
# Update data tables for more efficient computations&lt;br /&gt;
## Student would review and input additional data and IDs in the tables/sheets used to perform computations. This would be manual work (but super important), but would require high attention to detail.&lt;br /&gt;
## Additional Work: Requires Python/shell coding background. Student would run scripts that prepare and format data tables that are pushed to data.argosdb.org. Coding knowledge is needed in case of errors, bugs, or other mishaps in the code. Ongoing work as computations are performed.&lt;br /&gt;
# Curate and report on current pathogens to upload to ARGOS&lt;br /&gt;
## Student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found.&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
# QC Analysis using HIVE&lt;br /&gt;
## Analyze the curated pathogens using our QC ARGOS one-click pipeline.&lt;br /&gt;
## The results will be added to our ARGOS database.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the Fall.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Requirements for Completion ===&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; The following are mandatory. Failure to complete any will result in an incomplete volunteer record.&lt;br /&gt;
&lt;br /&gt;
==== Documentation ====&lt;br /&gt;
All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&lt;br /&gt;
&lt;br /&gt;
==== Written Report ====&lt;br /&gt;
Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&lt;br /&gt;
&lt;br /&gt;
==== Presentation &amp;amp; Slide Submission ====&lt;br /&gt;
Present your work last week of the 13-week period.&lt;br /&gt;
&lt;br /&gt;
Slides must be submitted to the POCs.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteers (TBD) ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Name&lt;br /&gt;
!Project Assigned&lt;br /&gt;
!POC Assigned&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
|Diya Kamalabharathy*&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod; Glyco web development&lt;br /&gt;
|-&lt;br /&gt;
|Anika Sikka&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|Akale Kinfe&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Nahom Abel&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Harivinay P. Gujjula&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|Sparsh Gupta&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Mathias Belay&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Isil Erbasol Serbes&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod, BiomarkerKB, ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|Ramtin Mashhoon&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|Anagha Kalle&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|Vishal Muthusekaran&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Adonay Awet&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Miao Wang*&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|ARGOS&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;nowiki&amp;gt;*&amp;lt;/nowiki&amp;gt;Returning volunteer.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=937</id>
		<title>Volunteership Fall 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=937"/>
		<updated>2025-08-20T18:15:48Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Volunteers (TBD) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== 2025 Volunteer Program Details ==&lt;br /&gt;
&lt;br /&gt;
=== Dates ===&lt;br /&gt;
&#039;&#039;&#039;Application Deadline&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 22, 2025, Noon (email your updated resume and projects in order of preference). Acceptance letter/email will be sent to candidates latest the day after the kick-off meeting.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Volunteer Zoom Kick-Off Meeting&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 25, 2025 | 4:00 to 5:00 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Program Dates: September 1st, 2025 – November 30th, 2025&#039;&#039;&#039; (13 weeks)&lt;br /&gt;
&lt;br /&gt;
Remote | Hybrid for GW employees and students (Ross Hall 5th floor)&lt;br /&gt;
&lt;br /&gt;
[https://hivelab.biochemistry.gwu.edu/wiki/Volunteership_2025 Summer 2025 Volunteership] (Closed)&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteer Expectations ===&lt;br /&gt;
&lt;br /&gt;
# Minimum commitment of 10 hours per week.&lt;br /&gt;
# Progress updates via Slack at least 3 days per week (scrum).&lt;br /&gt;
# Regular Zoom meetings with the assigned project point of contact.&lt;br /&gt;
# Attend some lectures or seminars remotely (max 4-5).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;Important:&#039;&#039;&#039; If the scrum is not updated for 2 consecutive working days, the candidate will be automatically dropped from the program.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Potential Projects ===&lt;br /&gt;
We are excited to continue our bioinformatics volunteership program in Fall 2025. This program offers students the opportunity to work on bioinformatics projects supported by agencies such as the NIH, ARPA-H, and FDA. Participants will gain exposure to a variety of activities within a bioinformatics lab, including data analysis, computational biology, and genomics. If you are interested, please email mazumder_lab@gwu.edu your resume and a ranked list of the projects that interest you most. You can also indicate if you want to focus on specific areas that are of interest to you.&lt;br /&gt;
# BiomarkerKB (biomarkerkb.org) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&lt;br /&gt;
# GlyGen (glygen.org) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information.&lt;br /&gt;
# ARGOS (argosdb.org) project: Analyze genomics data using HIVE to identify reference genome assemblies.&lt;br /&gt;
# PredictMod (hivelab.biochemistry.gwu.edu/predictmod) project. Curating PMIDs for intervention outcome prediction dataset LLM recommendation training.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
==== 1. BiomarkerKB Biocuration Project Ideas ====&lt;br /&gt;
POC: Daniall Masood, Maria Kim&lt;br /&gt;
&lt;br /&gt;
# Curate biomarkers for a specific disease&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
# Continue working on LLM methods started by volunteers over the summer.&lt;br /&gt;
## The data is available as well as some preliminary research and work done by previous volunteers in this area.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger, Urnisha Bhuiyan, Kate Warner&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer, Tianyi Wang, Pat McNeely (optional)&lt;br /&gt;
&lt;br /&gt;
Identifying relevant and useful publicly-available datasets for machine learning is currently a resource-intensive task. This curation project aims to develop a corpus for training an AI model to recommend PMIDs with publicly-available datasets useful for intervention outcome prediction models. The corpus will include an annotation spreadsheet + annotated PDFs for PubMed articles relevant to prostate, lung, breast, and liver cancer, and focus on indicators such as condition, intervention, and response.&lt;br /&gt;
&lt;br /&gt;
PMID curation involves:&lt;br /&gt;
&lt;br /&gt;
# Identify potentially relevant PMIDs that may have publicly-available datasets for training intervention outcome prediction models.&lt;br /&gt;
# Curate indicators of useful ML publications that could be used to train an LLM to recommend relevant publications for cancer modeling.&lt;br /&gt;
# Review peer curations and resolve annotation conflicts.&lt;br /&gt;
&lt;br /&gt;
Interested individuals should reach out to lorikrammer@gwu.edu.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
# Update data tables for more efficient computations&lt;br /&gt;
## Student would review and input additional data and IDs in the tables/sheets used to perform computations. This would be manual work (but super important), but would require high attention to detail.&lt;br /&gt;
## Additional Work: Requires Python/shell coding background. Student would run scripts that prepare and format data tables that are pushed to data.argosdb.org. Coding knowledge is needed in case of errors, bugs, or other mishaps in the code. Ongoing work as computations are performed.&lt;br /&gt;
# Curate and report on current pathogens to upload to ARGOS&lt;br /&gt;
## Student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found.&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
# QC Analysis using HIVE&lt;br /&gt;
## Analyze the curated pathogens using our QC ARGOS one-click pipeline.&lt;br /&gt;
## The results will be added to our ARGOS database.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Requirements for Completion ===&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; The following are mandatory. Failure to complete any will result in an incomplete volunteer record.&lt;br /&gt;
&lt;br /&gt;
==== Documentation ====&lt;br /&gt;
All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&lt;br /&gt;
&lt;br /&gt;
==== Written Report ====&lt;br /&gt;
Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&lt;br /&gt;
&lt;br /&gt;
==== Presentation &amp;amp; Slide Submission ====&lt;br /&gt;
Present your work last week of the 13-week period.&lt;br /&gt;
&lt;br /&gt;
Slides must be submitted to the POCs.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteers (TBD) ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Name&lt;br /&gt;
!Project Assigned&lt;br /&gt;
!POC Assigned&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
|Diya Kamalabharathy*&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod; Glyco web development&lt;br /&gt;
|-&lt;br /&gt;
|Anika Sikka&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|Akale Kinfe&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Nahom Abel&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Harivinay P. Gujjula&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|Sparsh Gupta&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Mathias Belay&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Isil Erbasol Serbes&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod, BiomarkerKB, ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|Ramtin Mashhoon&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|Anagha Kalle&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|Vishal Muthusekaran&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Adonay Awet&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|-&lt;br /&gt;
|Miao Wang*&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|ARGOS&lt;br /&gt;
|}&lt;br /&gt;
&amp;lt;nowiki&amp;gt;*&amp;lt;/nowiki&amp;gt;Returning volunteer.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=910</id>
		<title>Volunteership Fall 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_Fall_2025&amp;diff=910"/>
		<updated>2025-08-06T14:36:10Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Dates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== 2025 Volunteer Program Details ==&lt;br /&gt;
&lt;br /&gt;
=== Dates ===&lt;br /&gt;
&#039;&#039;&#039;Volunteer Zoom Kick-Off Meeting&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
August 25, 2025 | 4:00 to 5:00 PM&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Program Dates: September 1st, 2025 – November 30th, 2025&#039;&#039;&#039; (13 weeks)&lt;br /&gt;
&lt;br /&gt;
Remote&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteer Expectations ===&lt;br /&gt;
&lt;br /&gt;
# Minimum commitment of 10 hours per week.&lt;br /&gt;
# Progress updates via Slack at least 3 days per week (scrum).&lt;br /&gt;
# Regular Zoom meetings with the assigned project point of contact.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;Important:&#039;&#039;&#039; If the scrum is not updated for 2 consecutive working days, the candidate will be automatically dropped from the program.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Potential Projects ===&lt;br /&gt;
&lt;br /&gt;
# BiomarkerKB (biomarkerkb.org) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&lt;br /&gt;
# GlyGen (glygen.org) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information.&lt;br /&gt;
# ARGOS (argosdb.org) project: Analyze genomics data using HIVE to identify reference genome assemblies.&lt;br /&gt;
# PredictMod (hivelab.biochemistry.gwu.edu/predictmod) project. Curating PMIDs for intervention outcome prediction dataset LLM recommendation training.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
==== 1. BiomarkerKB Biocuration Project Ideas ====&lt;br /&gt;
POC: Daniall Masood, Maria Kim&lt;br /&gt;
&lt;br /&gt;
# Curate biomarkers for a specific disease (Alzheimers)&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer, Pat McNeeley&lt;br /&gt;
&lt;br /&gt;
PMID Curation:&lt;br /&gt;
&lt;br /&gt;
# Identify potentially relevant PMIDs that may have publicly-available datasets for training intervention outcome prediction models.&lt;br /&gt;
# Curate indicators of useful ML publications that could be used to train an LLM to recommend relevant publications for cancer modeling.&lt;br /&gt;
# Review peer curations and resolve annotation conflicts.&lt;br /&gt;
&lt;br /&gt;
Interested individuals should reach out to lorikrammer@gwu.edu.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
# Update data tables for more efficient computations&lt;br /&gt;
## Student would review and input additional data and IDs in the tables/sheets used to perform computations. This would be manual work (but super important), but would require high attention to detail.&lt;br /&gt;
## Additional Work: Requires Python/shell coding background. Student would run scripts that prepare and format data tables that are pushed to data.argosdb.org. Coding knowledge is needed in case of errors, bugs, or other mishaps in the code. Ongoing work as computations are performed.&lt;br /&gt;
# Curate and report on current pathogens to upload to ARGOS&lt;br /&gt;
## Student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found.&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
# QC Analysis using HIVE&lt;br /&gt;
## Analyze the curated pathogens using our QC ARGOS one-click pipeline.&lt;br /&gt;
## The results will be added to our ARGOS database.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Requirements for Completion ===&lt;br /&gt;
&#039;&#039;&#039;Note:&#039;&#039;&#039; The following are mandatory. Failure to complete any will result in an incomplete volunteer record.&lt;br /&gt;
&lt;br /&gt;
==== Documentation ====&lt;br /&gt;
All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&lt;br /&gt;
&lt;br /&gt;
==== Written Report ====&lt;br /&gt;
Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&lt;br /&gt;
&lt;br /&gt;
==== Presentation &amp;amp; Slide Submission ====&lt;br /&gt;
Present your work last week of the 13-week period.&lt;br /&gt;
&lt;br /&gt;
Slides must be submitted to the POCs.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Volunteers (TBD) ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Name&lt;br /&gt;
!Project&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;Anika Sikka (tentative)&#039;&#039;&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
#GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;Akale Kinfe&#039;&#039;&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Symposium_2025&amp;diff=880</id>
		<title>Symposium 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Symposium_2025&amp;diff=880"/>
		<updated>2025-07-17T19:15:40Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Agenda */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The HIVE Lab summer symposium is scheduled for Thursday July 31, 2025. It is an exciting time for the lab volunteers and interns to present their findings on the projects they worked on for 8 weeks.&lt;br /&gt;
&lt;br /&gt;
[[File:DC.png|center|frame]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;&#039;Program and Information&#039;&#039;&#039; ==&lt;br /&gt;
&lt;br /&gt;
=== &#039;&#039;&#039;Symposium Venue&#039;&#039;&#039; ===&lt;br /&gt;
The HIVE lab symposium will held in person at The George Washington University, Washington DC with an option to join virtually.&lt;br /&gt;
&lt;br /&gt;
In Person - Ross 647, Ross Hall, School of Health and Medical Sciences, The George Washington University, Washington DC ([https://maps.app.goo.gl/PHQmZacA4hWDvTCh6 MAP])&lt;br /&gt;
&lt;br /&gt;
Virtual - Zoom&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;&#039;Agenda&#039;&#039;&#039; ==&lt;br /&gt;
All times in Eastern Standard Time&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Time (ET)&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Project&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Title&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Presenter&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;10:00am&#039;&#039;&#039;&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |                                                                                            &#039;&#039;&#039;Welcome and Introduction&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Michael Tiemeyer (10 min)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
| colspan=&amp;quot;4&amp;quot; |                                                                                                                         &#039;&#039;Group 1 Moderator : Nathan Edwards&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|10:10am &lt;br /&gt;
|CFDE&lt;br /&gt;
|Integrating Biocuration and Data Standardization to Generate Machine Learning-Ready Glycan Datasets&lt;br /&gt;
|Ana Jaramillo and Yuxin Zou (20 min)&lt;br /&gt;
|-&lt;br /&gt;
|10:30am&lt;br /&gt;
|CFDE&lt;br /&gt;
|&lt;br /&gt;
|Campbell Ross (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|10:45am&lt;br /&gt;
|CFDE&lt;br /&gt;
|A Graph-Based AI Workflow for Mining Glycan Biomarkers and Related Annotations from Publications&lt;br /&gt;
|Cyrus Chun Hong Au Yeung (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|11:00am&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
|Sohana Bahl, Isaac Kim, Sparsh Gupta (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|11:15am&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
|Nathan Ressom, Ana Vohralikova, Mathias Belay (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|11:30am&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
|John McCaffery, Alma Ogunsina, Akale Kinfe (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;11:45am&#039;&#039;&#039;&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |&#039;&#039;&#039;Open Q and A&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;All (30 min)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|12:30pm&lt;br /&gt;
| colspan=&amp;quot;3&amp;quot; |                                                                                                          &#039;&#039;&#039;LUNCH (90 mins)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
| colspan=&amp;quot;4&amp;quot; |                                                                                                                         &#039;&#039;Group 1 Moderator : Nathan Edwards&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|2:00pm&lt;br /&gt;
|Predictmod AI-READI&lt;br /&gt;
|Robust Classification of Glycemic Health States from Continuous Glucose &lt;br /&gt;
|Nikhil Arethiya (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|2:15pm&lt;br /&gt;
|Predictmod Curation&lt;br /&gt;
|PredictMod: PubMed Curation for Training an LLM for Recommendation&lt;br /&gt;
|Grace Chong, Aaron Ressom, Diya Kamalabharathy (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|2:30pm&lt;br /&gt;
|Argos&lt;br /&gt;
|Curation of Emerging Pathogen Genomes for FDA-ARGOS Database Expansion&lt;br /&gt;
|Miao Wang (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|2:45pm&lt;br /&gt;
|GlyGen&lt;br /&gt;
|GlyGen Biocuration Project&lt;br /&gt;
|Aise Arpinar, Haravinay P. Gujjulla, Nahom Abel (20 min)&lt;br /&gt;
|-&lt;br /&gt;
|3:05pm&lt;br /&gt;
|GlycoSiteMiner&lt;br /&gt;
|&lt;br /&gt;
|(15 min)&lt;br /&gt;
|-&lt;br /&gt;
|3:20pm&lt;br /&gt;
|Glycobiology Web Development&lt;br /&gt;
|A Resource Drill Down and Visualization for the Glyspace Alliance&lt;br /&gt;
|Diya Kamalabharathy (5 min)&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;3:25pm&#039;&#039;&#039;&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |&#039;&#039;&#039;Open Q and A&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;All (20 min)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|3:45pm&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |                                                                                       &#039;&#039;&#039;Closing Remarks&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Raja Mazumder&#039;&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;&#039;Project Description&#039;&#039;&#039; ==&lt;br /&gt;
&lt;br /&gt;
=== GlyGen Project ===&lt;br /&gt;
The GlyGen Biocuration project focuses on integrating legacy, yet valuable, data from the CarbBank and CFG databases into the GlyGen infrastructure. A key challenge is mapping metadata, such as species names and publication references, to standardized dictionaries and ontologies. While most entries have been automatically matched using custom scripts, remaining inconsistencies, including outdated, misspelled, or abbreviated terms, require manual curation using resources such as Google, PubMed, and domain-specific dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
=== BiomarkerKB Biocuration Project ===&lt;br /&gt;
The Biomarker Biocuration project focuses on biomarker curation from abstracts and publications in the BiomarkerKB data model. A key challenge in curating biomarkers is the vast amount of data that is present over various publications. Manual curation requires reading, inferring, and understanding key elements of biomarker data and being able to map it to the defined biomarker data model. LLM methodologies will help immensely in being able to recognize biomarker and condition data and being able to map information found into the data model while also automatically mapping other contextual and standardized data to the model to allow data to be AI andmachine leanring ready.&lt;br /&gt;
&lt;br /&gt;
=== ArgosDB Curation Project ===&lt;br /&gt;
This project focuses on evaluating and curating high-quality genomes of emerging and clinically relevant pathogens, with an emphasis on fungal species. Using public genomic repositories and FDA-ARGOS inclusion criteria, I identify candidate organisms for database expansion to support diagnostic assay development and public health surveillance.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Symposium_2025&amp;diff=879</id>
		<title>Symposium 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Symposium_2025&amp;diff=879"/>
		<updated>2025-07-17T18:47:19Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Project Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The HIVE Lab summer symposium is scheduled for Thursday July 31, 2025. It is an exciting time for the lab volunteers and interns to present their findings on the projects they worked on for 8 weeks.&lt;br /&gt;
&lt;br /&gt;
[[File:DC.png|center|frame]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;&#039;Program and Information&#039;&#039;&#039; ==&lt;br /&gt;
&lt;br /&gt;
=== &#039;&#039;&#039;Symposium Venue&#039;&#039;&#039; ===&lt;br /&gt;
The HIVE lab symposium will held in person at The George Washington University, Washington DC with an option to join virtually.&lt;br /&gt;
&lt;br /&gt;
In Person - Ross 647, Ross Hall, School of Health and Medical Sciences, The George Washington University, Washington DC ([https://maps.app.goo.gl/PHQmZacA4hWDvTCh6 MAP])&lt;br /&gt;
&lt;br /&gt;
Virtual - Zoom&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;&#039;Agenda&#039;&#039;&#039; ==&lt;br /&gt;
All times in Eastern Standard Time&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Time (ET)&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Project&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Title&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Presenter&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;10:00am&#039;&#039;&#039;&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |                                                                                            &#039;&#039;&#039;Welcome and Introduction&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Michael Tiemeyer (10 min)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
| colspan=&amp;quot;4&amp;quot; |                                                                                                                         &#039;&#039;Group 1 Moderator : Nathan Edwards&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|10:10am &lt;br /&gt;
|CFDE&lt;br /&gt;
|Integrating Biocuration and Data Standardization to Generate Machine Learning-Ready Glycan Datasets&lt;br /&gt;
|Ana Jaramillo and Yuxin Zou (20 min)&lt;br /&gt;
|-&lt;br /&gt;
|10:30am&lt;br /&gt;
|CFDE&lt;br /&gt;
|&lt;br /&gt;
|Campbell Ross (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|10:45am&lt;br /&gt;
|CFDE&lt;br /&gt;
|A Graph-Based AI Workflow for Mining Glycan Biomarkers and Related Annotations from Publications&lt;br /&gt;
|Cyrus Chun Hong Au Yeung (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|11:00am&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
|Sohana Bahl, Isaac Kim, Sparsh Gupta (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|11:15am&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
|Nathan Ressom, Ana Vohralikova, Mathias Belay (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|11:30am&lt;br /&gt;
|BiomarkerKB&lt;br /&gt;
|&lt;br /&gt;
|John McCaffery, Alma Ogunsina, Akale Kinfe (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;11:45am&#039;&#039;&#039;&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |&#039;&#039;&#039;Open Q and A&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;All (30 min)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|12:30pm&lt;br /&gt;
| colspan=&amp;quot;3&amp;quot; |                                                                                                          &#039;&#039;&#039;LUNCH (90 mins)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
| colspan=&amp;quot;4&amp;quot; |                                                                                                                         &#039;&#039;Group 1 Moderator : Nathan Edwards&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|2:00pm&lt;br /&gt;
|Predictmod AI-READI&lt;br /&gt;
|Robust Classification of Glycemic Health States from Continuous Glucose &lt;br /&gt;
|Nikhil Arethiya (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|2:15pm&lt;br /&gt;
|Predictmod Curation&lt;br /&gt;
|PredictMod: PubMed Curation for Training an LLM for Recommendation&lt;br /&gt;
|Grace Chong, Aaron Ressom, Diya Kamalabharathy (15 min)&lt;br /&gt;
|-&lt;br /&gt;
|2:30pm&lt;br /&gt;
|Argos&lt;br /&gt;
|Curation of Emerging Pathogen Genomes for FDA-ARGOS Database Expansion&lt;br /&gt;
|(15 min)&lt;br /&gt;
|-&lt;br /&gt;
|2:45pm&lt;br /&gt;
|GlyGen&lt;br /&gt;
|GlyGen Biocuration Project&lt;br /&gt;
|Aise Arpinar, Haravinay P. Gujjulla, Nahom Abel (20 min)&lt;br /&gt;
|-&lt;br /&gt;
|3:05pm&lt;br /&gt;
|GlycoSiteMiner&lt;br /&gt;
|&lt;br /&gt;
|(15 min)&lt;br /&gt;
|-&lt;br /&gt;
|3:20pm&lt;br /&gt;
|Glycobiology Web Development&lt;br /&gt;
|A Resource Drill Down and Visualization for the Glyspace Alliance&lt;br /&gt;
|Diya Kamalabharathy (5 min)&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;3:25pm&#039;&#039;&#039;&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |&#039;&#039;&#039;Open Q and A&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;All (20 min)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|3:45pm&lt;br /&gt;
| colspan=&amp;quot;2&amp;quot; |                                                                                       &#039;&#039;&#039;Closing Remarks&#039;&#039;&#039;&lt;br /&gt;
|&#039;&#039;&#039;Raja Mazumder&#039;&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;&#039;Project Description&#039;&#039;&#039; ==&lt;br /&gt;
&lt;br /&gt;
=== GlyGen Project ===&lt;br /&gt;
The GlyGen Biocuration project focuses on integrating legacy, yet valuable, data from the CarbBank and CFG databases into the GlyGen infrastructure. A key challenge is mapping metadata, such as species names and publication references, to standardized dictionaries and ontologies. While most entries have been automatically matched using custom scripts, remaining inconsistencies, including outdated, misspelled, or abbreviated terms, require manual curation using resources such as Google, PubMed, and domain-specific dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
=== BiomarkerKB Biocuration Project ===&lt;br /&gt;
The Biomarker Biocuration project focuses on biomarker curation from abstracts and publications in the BiomarkerKB data model. A key challenge in curating biomarkers is the vast amount of data that is present over various publications. Manual curation requires reading, inferring, and understanding key elements of biomarker data and being able to map it to the defined biomarker data model. LLM methodologies will help immensely in being able to recognize biomarker and condition data and being able to map information found into the data model while also automatically mapping other contextual and standardized data to the model to allow data to be AI andmachine leanring ready.&lt;br /&gt;
&lt;br /&gt;
=== ArgosDB Curation Project ===&lt;br /&gt;
This project focuses on evaluating and curating high-quality genomes of emerging and clinically relevant pathogens, with an emphasis on fungal species. Using public genomic repositories and FDA-ARGOS inclusion criteria, I identify candidate organisms for database expansion to support diagnostic assay development and public health surveillance.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Additional_ARGOS_Reviewed_Organisms&amp;diff=856</id>
		<title>Additional ARGOS Reviewed Organisms</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Additional_ARGOS_Reviewed_Organisms&amp;diff=856"/>
		<updated>2025-06-05T16:15:16Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Back to the [[FDA-ARGOS WIKI|Home Page]] for FDA-ARGOS&lt;br /&gt;
&lt;br /&gt;
This page highlights organisms that were selected, reviewed, and added to ARGOSDB that are external to the FDA-ARGOS BioProject PRJNA231221. The QC and metadata information for these organisms can be found in these tables on data.argosdb.org:&lt;br /&gt;
&lt;br /&gt;
- assemblyQC_ARGOS&lt;br /&gt;
&lt;br /&gt;
- biosampleMeta_ARGOS&lt;br /&gt;
&lt;br /&gt;
- ngsQC_ARGOS&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Organism List==&lt;br /&gt;
The organisms listed below can also be found on the [https://data.argosdb.org/ARGOS_000018 ngs_id_list.tsv table] on data.argosdb.org.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Organisms that have been QC&#039;d and reviewed:&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
- Influenza A (H3N2) 2022&lt;br /&gt;
&lt;br /&gt;
- Influenza A (H1N1) 2022&lt;br /&gt;
&lt;br /&gt;
- Influenza A (Puerto Rico H1N1) reference genome&lt;br /&gt;
&lt;br /&gt;
- Monkeypox Virus (WRAIR 761)&lt;br /&gt;
&lt;br /&gt;
- Monkeypox Virus (USA2003(&lt;br /&gt;
&lt;br /&gt;
- Human immunodeficiency virus type 1&lt;br /&gt;
&lt;br /&gt;
- Orthomarburgvirus marburgense (isolate Ravn virus/H.sapiens-tc/KEN/1987/Kitum Cave-810040)&lt;br /&gt;
&lt;br /&gt;
- Lake Victoria marburgvirus (Angola2005 strain Ang1379c)&lt;br /&gt;
&lt;br /&gt;
- Marburg virus (Musoke Kenya 1980 for SAMN11077998)&lt;br /&gt;
&lt;br /&gt;
- Marburg virus (Musoke Kenya 1980 for SAMN16357613)&lt;br /&gt;
&lt;br /&gt;
- Lake Victoria marburgvirus (Ci67)&lt;br /&gt;
&lt;br /&gt;
- Sudan ebolavirus (Gulu)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Omicron/XBB.1.5 SARS-CoV-2/human/USA/CA-CDC-STM-S85G6U7MH/2023)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2/human/USA/WA-CDC-02982586-001/2020)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (isolate Wuhan-Hu-1)&lt;br /&gt;
&lt;br /&gt;
- Candidozyma auris&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Organisms that were selected, but have not been QC&#039;d due to missing data information:&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
- Middle East respiratory syndrome related coronavirus (EMC/2012)&lt;br /&gt;
&lt;br /&gt;
- Middle East respiratory syndrome related coronavirus (VR-3248SD)&lt;br /&gt;
&lt;br /&gt;
- Middle East respiratory syndrome related coronavirus (IRF0021 MERS JOR)&lt;br /&gt;
&lt;br /&gt;
- Middle East respiratory syndrome related coronavirus (IRF0038 MERS EMC)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Delta/B.1.617.2)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Omicron/B.1.1.529)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Beta/B.1.351)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Alpha/B.1.1.7)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Gamma/P.1)&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_FAQs&amp;diff=855</id>
		<title>FDA-ARGOS FAQs</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_FAQs&amp;diff=855"/>
		<updated>2025-06-05T16:13:32Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
Back to [[FDA-ARGOS WIKI|Home Page]] for FDA-ARGOS&lt;br /&gt;
&lt;br /&gt;
===== What is the ArgosDB and how is it organized? =====&lt;br /&gt;
ArgosDB was developed as a result of expanded funding for the FDA-ARGOS project, which is described in detail in the [[About Argos DataBase|About]] page of this website. The database stores cross-kingdom QC attributes of clinically relevant organisms organized into respective datasets. The current datasets (as of March 2025) are: ngsQC_ARGOS, ngsQC_ARGOS_unreviewed, assemblyQC_ARGOS, assemblyQC_ARGOS_unreviewed, biosampleMeta_ARGOS, and biosampleMeta_ARGOS unreviewed. The original four key datasets are ngsQC_HIVE, assemblyQC_HIVE, siteQC_HIVE, and biosampleMeta_HIVE. These datasets are associated with core QC protocols, which are documented via Biocompute Objects (BCO) and organized under their BCO IDs. &lt;br /&gt;
&lt;br /&gt;
When new QC data has been produced per an organism of interest, the respective dataset(s) is/are appended, which is documented per data release in the FDA-ARGOS GitHub (&amp;lt;nowiki&amp;gt;https://github.com/FDA-ARGOS&amp;lt;/nowiki&amp;gt;). All datasets are in alignment with the current data dictionary (v1.6 as of March 2025), which guides the QC process for that dataset as well as the column headers for a given dataset. Datasets are available as either .tsv or FASTA files if associated with a genome assembly, and all datasets and BCOs are available for download. All data provenance and curation are captured and reproducible via their BCO. Additional datasets include data from the original FDA BioProject, the Data Dictionary, Drug Resistance Mutations, Genome Assemblies (multiple), and a mapping key that assists in linking all the available data via important accessions. &lt;br /&gt;
&lt;br /&gt;
A total of 24 datasets are available as of 03/2025.&lt;br /&gt;
&lt;br /&gt;
===== How can I view or access the previous versions of the data? =====&lt;br /&gt;
First, go to the Release History tab on the ARGOS home page. Next, click details on the desired data object. Then select the desired version and data from the version transition dropbox on the top left corner and view metrics such as field count, fields added, fields removed, row count, row count prev, rows count change, ID count, IDs added, and IDs removed.&lt;br /&gt;
&lt;br /&gt;
===== What does the schema version in the datasets refer to? =====&lt;br /&gt;
The schema relates to the organization of the ARGOS data within the data model. The version is reflective of the FDA-ARGOS data dictionary version that is currently applied to all updated datasets. As of March 2025, the current schema is v1.6 and can be found in the [https://github.com/FDA-ARGOS/data.argosdb/tree/main/data_dictionary/v1.6 FDA-ARGOs GitHub]. &lt;br /&gt;
&lt;br /&gt;
===== Does Argos have a tutorial on how to use the site? =====&lt;br /&gt;
Yes! Please follow the below basic instructions of how to navigate the DB:  &lt;br /&gt;
&lt;br /&gt;
====== How to find and search a dataset ======&lt;br /&gt;
On data.argosdb.org home page, you can search a dataset by entering the keyword in Search Datasets. &lt;br /&gt;
&lt;br /&gt;
Keywords can be BCO ID, organism name, or even a term that describes biological processes. In the following example, three results appear upon the search for ebola.  &lt;br /&gt;
To further narrow down the result, select filters on the left side bar. Alternatively, users can search datasets by selecting relevant filters on the left side bar.[[File:Screenshot 2025-03-06 at 1.55.51 PM.png|thumb|719x719px|data.argosdb Home Page|none]][[File:Ebola search.png|thumb|722x722px|&#039;ebola&#039; searched in the search bar of the ARGOS database|none]]&lt;br /&gt;
&lt;br /&gt;
====== How to select a dataset ======&lt;br /&gt;
Next, to select a dataset, click on view details under DETAILS. Previous released versioned datasets are available upon clicking the dropdown button &lt;br /&gt;
[[File:Dataset.png|thumb|722x722px|view of an example dataset after clicking on the &#039;...view details&#039; link on the homepage. The dropdown menu at the top lets you select data versions.|none]] &lt;br /&gt;
&lt;br /&gt;
====== How to view and download the BCO that is corresponding to its dataset ======&lt;br /&gt;
To download the dataset, click on the DOWNLOADS tab and select the download format for the target dataset. BCO JSON will be downloaded and automatically opened as a .txt file upon clicking on Download BCO. Dataset will be downloaded and automatically opened either as .tsv or .csv file upon clicking on Download dataset file.&lt;br /&gt;
[[File:Bco.png|thumb|720x720px|BCO JSON tab of the dataset.|none]]&lt;br /&gt;
[[File:Screenshot 2025-03-06 at 2.08.02 PM.png|thumb|724x724px|Downloads tab for the dataset. BCO and table can be downloaded here.|none]]&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Additional_ARGOS_Reviewed_Organisms&amp;diff=854</id>
		<title>Additional ARGOS Reviewed Organisms</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Additional_ARGOS_Reviewed_Organisms&amp;diff=854"/>
		<updated>2025-06-02T15:25:39Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page highlights organisms that were selected, reviewed, and added to ARGOSDB that are external to the FDA-ARGOS BioProject PRJNA231221. The QC and metadata information for these organisms can be found in these tables on data.argosdb.org:&lt;br /&gt;
&lt;br /&gt;
- assemblyQC_ARGOS&lt;br /&gt;
&lt;br /&gt;
- biosampleMeta_ARGOS&lt;br /&gt;
&lt;br /&gt;
- ngsQC_ARGOS&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Organism List==&lt;br /&gt;
The organisms listed below can also be found on the [https://data.argosdb.org/ARGOS_000018 ngs_id_list.tsv table] on data.argosdb.org.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Organisms that have been QC&#039;d and reviewed:&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
- Influenza A (H3N2) 2022&lt;br /&gt;
&lt;br /&gt;
- Influenza A (H1N1) 2022&lt;br /&gt;
&lt;br /&gt;
- Influenza A (Puerto Rico H1N1) reference genome&lt;br /&gt;
&lt;br /&gt;
- Monkeypox Virus (WRAIR 761)&lt;br /&gt;
&lt;br /&gt;
- Monkeypox Virus (USA2003(&lt;br /&gt;
&lt;br /&gt;
- Human immunodeficiency virus type 1&lt;br /&gt;
&lt;br /&gt;
- Orthomarburgvirus marburgense (isolate Ravn virus/H.sapiens-tc/KEN/1987/Kitum Cave-810040)&lt;br /&gt;
&lt;br /&gt;
- Lake Victoria marburgvirus (Angola2005 strain Ang1379c)&lt;br /&gt;
&lt;br /&gt;
- Marburg virus (Musoke Kenya 1980 for SAMN11077998)&lt;br /&gt;
&lt;br /&gt;
- Marburg virus (Musoke Kenya 1980 for SAMN16357613)&lt;br /&gt;
&lt;br /&gt;
- Lake Victoria marburgvirus (Ci67)&lt;br /&gt;
&lt;br /&gt;
- Sudan ebolavirus (Gulu)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Omicron/XBB.1.5 SARS-CoV-2/human/USA/CA-CDC-STM-S85G6U7MH/2023)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2/human/USA/WA-CDC-02982586-001/2020)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (isolate Wuhan-Hu-1)&lt;br /&gt;
&lt;br /&gt;
- Candidozyma auris&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Organisms that were selected, but have not been QC&#039;d due to missing data information:&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
- Middle East respiratory syndrome related coronavirus (EMC/2012)&lt;br /&gt;
&lt;br /&gt;
- Middle East respiratory syndrome related coronavirus (VR-3248SD)&lt;br /&gt;
&lt;br /&gt;
- Middle East respiratory syndrome related coronavirus (IRF0021 MERS JOR)&lt;br /&gt;
&lt;br /&gt;
- Middle East respiratory syndrome related coronavirus (IRF0038 MERS EMC)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Delta/B.1.617.2)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Omicron/B.1.1.529)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Beta/B.1.351)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Alpha/B.1.1.7)&lt;br /&gt;
&lt;br /&gt;
- Severe acute respiratory syndrome coronavirus 2 (Gamma/P.1)&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=842</id>
		<title>FDA-ARGOS WIKI</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=FDA-ARGOS_WIKI&amp;diff=842"/>
		<updated>2025-05-14T15:53:08Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* FDA-ARGOS Database */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
FDA-ARGOS database updates may help researchers rapidly validate diagnostic tests and use qualified genetic sequences to support future product development. NCBI BioProject [https://www.ncbi.nlm.nih.gov/bioproject/231221 PRJNA231221]&lt;br /&gt;
&lt;br /&gt;
As of September 2021, Embleema and George Washington University have been conducting bioinformatic research and system development, focusing on expanding the FDA-ARGOS database. This project expands datasets publicly available in FDA-ARGOS, improves quality control by developing quality matrix tools and scoring approaches that will allow the mining of public sequence databases, and identifies high-quality sequences for upload to the FDA-ARGOS database as regulatory-grade sequences. Building on expansions during the COVID-19 pandemic, this project aims to further improve the utility of the FDA-ARGOS database as a key tool for medical countermeasure development and validation.&lt;br /&gt;
&lt;br /&gt;
For additional details on project information and assembly QC see&lt;br /&gt;
&lt;br /&gt;
* [https://www.fda.gov/emergency-preparedness-and-response/preparedness-research/expanding-next-generation-sequencing-tools-support-pandemic-preparedness-and-response FDA-ARGOS Project Information]&lt;br /&gt;
* [https://data.argosdb.org/ ARGOS Database]&lt;br /&gt;
* [[About Argos DataBase]]&lt;br /&gt;
* [https://www.fda.gov/medical-devices/science-and-research-medical-devices/database-reference-grade-microbial-sequences-fda-argos FDA Statement for FDA-ARGOS Database]&lt;br /&gt;
* [[ARGOS Contact Us]]&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;FDA-ARGOS Initial Phase&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In May 2014, the FDA and collaborators had established a publicly available database for Reference Grade microbial Sequences called FDA-ARGOS. With funding support from FDA&#039;s Office of Counterterrorism and Emerging Threats (OCET) and DoD, the FDA-ARGOS team had initially collected and sequenced 2000 microbes that included biothreat microorganisms, common clinical pathogens, and closely related species. At the beginning of this project, the FDA-ARGOS microbial genomes were generated in 3 phases. Generally:&lt;br /&gt;
&lt;br /&gt;
* Phase 1, entailed collection of a previously identified microbe and nucleic acid extraction.&lt;br /&gt;
* Phase 2, the microbial nucleic acids were then sequenced and de novo assembled occurred using Illumina and PacBio sequencing platforms at the Institute for Genome Sciences at the University of Maryland (UMD-IGS).&lt;br /&gt;
* Phase 3, the assembled genomes were then vetted by an ID-NGS subject matter expert working group that consisted of FDA personnel and collaborators and the data was then deposited in the NCBI databases.&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS genomes meet the quality metrics for reference-grade genomes for regulatory use. FDA-ARGOS reference genomes have been de novo assembled with high depth of base coverage and placed within a pre-established phylogenetic tree. Each microbial isolate in the database is covered at a minimum of 20X over 95 percent of the assembled core genome. Furthermore, sample-specific metadata, raw reads, assemblies, annotation, and details of the bioinformatics pipeline are available.&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS Database ==&lt;br /&gt;
&lt;br /&gt;
The FDA-ARGOS database ([https://data.argosdb.org/ data.argosdb.org]) is a public database containing quality-controlled reference genomes for diagnostic and regulatory purposes.&lt;br /&gt;
&lt;br /&gt;
* FDA-ARGOS NCBI BioProject can be found [https://www.ncbi.nlm.nih.gov/bioproject/231221 here].&lt;br /&gt;
* [[Additional ARGOS Reviewed Organisms]].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a visual display of data analytics, view the prototype dashboard [https://studyanalytics.embleema.com/superset/dashboard/argos/?standalone=2&amp;amp;expand_filters=0 here].&lt;br /&gt;
&lt;br /&gt;
== Project Publications ==&lt;br /&gt;
&lt;br /&gt;
*Sichtig, H., Minogue, T., Yan, Y. &#039;&#039;et al.&#039;&#039; FDA-ARGOS is a database with public quality-controlled reference genomes for diagnostic use and regulatory science. &#039;&#039;Nat Commun&#039;&#039; 10, 3313 (2019). https://doi.org/10.1038/s41467-019-11306-6&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== FDA-ARGOS FAQs ==&lt;br /&gt;
Frequently asked questions about FDA-ARGOS can be found [[FDA-ARGOS FAQs|here]]. If there are any further questions, feel free to [[ARGOS Contact Us|contact us]].&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_2025&amp;diff=839</id>
		<title>Volunteership 2025</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=Volunteership_2025&amp;diff=839"/>
		<updated>2025-05-13T15:48:11Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;h2&amp;gt;2025 Volunteer Program Details&amp;lt;/h2&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h3&amp;gt;Dates&amp;lt;/h3&amp;gt;&lt;br /&gt;
&amp;lt;strong&amp;gt;Volunteer Zoom Kick-Off Meeting&amp;lt;/strong&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
May 26, 2025 | 3:30 to 4:30 PM&lt;br /&gt;
&lt;br /&gt;
&amp;lt;strong&amp;gt;Program Dates: June 2nd, 2025 – July 25th, 2025&amp;lt;/strong&amp;gt; (8 weeks)&amp;lt;br&amp;gt;&lt;br /&gt;
Monday to Friday | Remote | No breaks&lt;br /&gt;
&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h3&amp;gt;Volunteer Expectations&amp;lt;/h3&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;Daily progress updates via Slack (scrum).&amp;lt;/li&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;Regular Zoom meetings with the assigned project point of contact.&amp;lt;/li&amp;gt;&amp;lt;li&amp;gt;Expected to dedicate 5–6 hours per day to project work, with the remaining time focused on skill development or reading. &amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;p style=&amp;quot;color: red;&amp;quot;&amp;gt;&amp;lt;strong&amp;gt;Important:&amp;lt;/strong&amp;gt; If the scrum is not updated for 2 consecutive days, the candidate will be &amp;lt;u&amp;gt;automatically dropped&amp;lt;/u&amp;gt; from the program.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h3&amp;gt;Potential Projects&amp;lt;/h3&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;BiomarkerKB ([https://biomarkerkb.org biomarkerkb.org]) project: Biomarker curation project. Involves reading papers and collecting biomarkers.&amp;lt;/li&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;GlyGen ([https://glygen.org glygen.org]) project: Review glycomics and glycoproteomics data and curate tissue, disease, and other related information. &amp;lt;/li&amp;gt;&amp;lt;li&amp;gt;ARGOS ([https://argosdb.org argosdb.org]) project: Analyze genomics data using HIVE to identify reference genome assemblies. &amp;lt;/li&amp;gt;&amp;lt;li&amp;gt;PredictMod ([https://hivelab.biochemistry.gwu.edu/predictmod hivelab.biochemistry.gwu.edu/predictmod]) project. Identifying datasets and harmonizing them so that they can be used to generate ML models.  &amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&#039;&#039;Note: Individuals involved in the above projects with a background in programming and/or machine learning may also undertake additional tasks to support the development of ML models, which can be integrated into PredictMod or used to enhance AI/ML-ready datasets within GlyGen.&#039;&#039;&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h4&amp;gt;1. BiomarkerKB Biocuration Project Ideas&amp;lt;/h4&amp;gt;POC: Daniall Masood, Maria Kim&lt;br /&gt;
# Curate biomarkers for a specific disease (Alzheimers)&lt;br /&gt;
## The student would be doing manual curation for about 4 weeks, with regular check-ins with me to ensure it is being done correctly.&lt;br /&gt;
## The next 4 weeks can be dedicated to developing an LLM or an automated process to extract biomarker details with data collected in the first 4 weeks as training data/example data.&lt;br /&gt;
# Top 50 biomarkers&lt;br /&gt;
## Curate the top 50 biomarkers for biomarkerkb.org.&lt;br /&gt;
## Define what constitutes a top 50 biomarker.&lt;br /&gt;
## Begin curating biomarkers from different sources and papers by collecting fields mentioned in the data model, as well as collecting cross-references.&lt;br /&gt;
# Biocuration of biomarkers from NLP/LLM work&lt;br /&gt;
## Use the biomarkers collected from NLP work.&lt;br /&gt;
## Curate biomarkers. Data provided was not provided in the biomarker data model.&lt;br /&gt;
## While curating the biomarkers, check if data collected from NLP is correct.&lt;br /&gt;
## After completion, the student can start using curated data to work on the NLP/LLM method.&lt;br /&gt;
# Curate biomarkers for a treatment&lt;br /&gt;
## See #1 above.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas, diseases, treatments, or methods they want to focus on, please reach out to daniallmasood@gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&lt;br /&gt;
&lt;br /&gt;
==== 2. GlyGen Biocuration Project Ideas ====&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
Over the last three decades, numerous glycomics database projects have been initiated to collect valuable information about glycans, proteins, and their interactions. Some of these databases have been discontinued due to the end of project funding. However, the data within these databases remains highly valuable to the community. Integrating these datasets into modern databases or knowledgebases, such as GlyGen, presents a challenge because much of the valuable metadata (e.g., species, tissue, disease, cell line) annotations are free-text terms that do not align with established standard dictionaries and ontologies used in modern resources. Automated matching of this information with dictionaries or ontologies is often not possible due to the use of synonyms, spelling errors, or abbreviations. For example, &amp;quot;human,&amp;quot; &amp;quot;man,&amp;quot; and &amp;quot;h. sapiens&amp;quot; all map to the scientific species name &amp;quot;Homo sapiens.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
The GlyGen project aims to make datasets from two older databases (CarbBank, CFG) accessible by migrating the data and metadata into our database. For this project, we are seeking curators with a medical or biology background who are interested in helping map metadata terms from these old databases to standard dictionaries and ontologies.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using internet resources (e.g., Google, Wikipedia) to identify terms used in the old database.&lt;br /&gt;
# Mapping identified terms to corresponding dictionaries and ontologies using the webpages and search interfaces of these projects.&lt;br /&gt;
# Finding papers based on titles and author lists that may contain spelling errors.&lt;br /&gt;
# Interacting and discussing with other curators in case terms are mapped differently.&lt;br /&gt;
&lt;br /&gt;
If you have any other ideas or methods you would like to focus on, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;3. GlyGen Publication Analysis Project Ideas&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Rene Ranzinger and Urnisha Bhuiyan&lt;br /&gt;
&lt;br /&gt;
One of the challenges for any bioinformatics project is understanding the size of its community, how well the project serves this community, and how widely its software/database is used. A potential solution is to analyze PubMed publication data. We are seeking applicants with programming skills (in Python or Java) to perform this analysis.&lt;br /&gt;
&lt;br /&gt;
The project involves:&lt;br /&gt;
&lt;br /&gt;
# Using the PubMed web API to filter publications based on keywords.&lt;br /&gt;
# Analyzing paper abstracts to identify research institutions and groups that form the community.&lt;br /&gt;
# Filtering the community list to exclude unrelated co-authors.&lt;br /&gt;
# Prioritize papers identified by GlycoSiteMiner for curation via TableMaker&lt;br /&gt;
&lt;br /&gt;
A subproject will involve analyzing the full text of papers (when available) for keywords or resource and database names. The results of the analysis will be discussed with GlyGen project member who will suggest changes and improvements to the analysis and data presentation. Source code developed as part of this project will be documented and shared in a public GitHub repository. If you have any other ideas or methods you would like to explore, please reach out to rene@ccrc.uga.edu to discuss them.&lt;br /&gt;
&lt;br /&gt;
==== 4. PredictMod Machine Learning Project Ideas ====&lt;br /&gt;
POC: Lori Krammer&lt;br /&gt;
&lt;br /&gt;
Data Identification &amp;amp; Harmonization: &lt;br /&gt;
&lt;br /&gt;
# Identify publicly-available datasets from scientific literature that can be used for intervention outcome prediction models.&lt;br /&gt;
# Conduct data harmonization and pre-processing following established project pipelines to make ML-ready dataset and data dictionary.&lt;br /&gt;
&lt;br /&gt;
Modeling &amp;amp; Integration (for those with experience in programming/ML)&lt;br /&gt;
&lt;br /&gt;
# Perform model training and document ML pipeline in a BioCompute Object (BCO). &lt;br /&gt;
# Integrate model into PredictMod platform.&lt;br /&gt;
&lt;br /&gt;
Individuals with a background or interest in machine learning should reach out to lorikrammer@gwu.edu with a potential dataset to determine if it is a feasible project for the summer.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;5. FDA-ARGOS Computation and Pathogen Curation Project&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
POC: Christie Woodside, Jonathon Keeney&lt;br /&gt;
&lt;br /&gt;
# Update data tables for more efficient computations&lt;br /&gt;
## Student would review and input additional data and IDs in the tables/sheets used to perform computations. This would be manual work (but super important), but would require high attention to detail. ~1 week&#039;s worth of work&lt;br /&gt;
## Requires Python/shell coding background. Student would run scripts that prepare and format data tables that are pushed to data.argosdb.org. Coding knowledge is needed in case of errors, bugs, or other mishaps in the code. Ongoing work as computations are performed.&lt;br /&gt;
# Curate and report on current pathogens to upload to ARGOS&lt;br /&gt;
## Student would work on manual curation of circulating pathogens to be added to data.argosdb.org. Regular check-ins and reports of what was found. ~4-10 weeks worth of work&lt;br /&gt;
## Locate assembly IDs, reads, and metagenomic information for these pathogens to be used in computations and deposited into data.argosdb.org.&lt;br /&gt;
## Provide documentation on why they were curated, why they are important, how they were selected, and how data was collected.&lt;br /&gt;
&lt;br /&gt;
If the student has any other ideas or methods they want to focus on, please reach out to christie.woodside@email.gwu.edu to discuss your idea and check if it will be feasible as a project for the summer.&amp;lt;hr&amp;gt;&lt;br /&gt;
&amp;lt;h3&amp;gt;Requirements for Completion&amp;lt;/h3&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;strong&amp;gt;Note:&amp;lt;/strong&amp;gt; The following are &amp;lt;u&amp;gt;mandatory&amp;lt;/u&amp;gt;. Failure to complete any will result in an incomplete volunteer record.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h4&amp;gt;Documentation&amp;lt;/h4&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;All volunteers must maintain adequate documentation of their work, including written protocols and scripts submitted to GitHub.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h4&amp;gt;Written Report&amp;lt;/h4&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Submit a 1–2 page summary of your tasks and accomplishments to the Admin during the final week of your program.&amp;lt;/p&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;h4&amp;gt;Presentation &amp;amp; Slide Submission&amp;lt;/h4&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Present your work last week of the 8-week period.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Slides must be submitted to the Admin Team and should include:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;A title slide with your name, date, and mentor&amp;lt;/li&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;At least 3 content slides&amp;lt;/li&amp;gt;&lt;br /&gt;
  &amp;lt;li&amp;gt;A final slide with acknowledgements or references&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ul&amp;gt;&lt;br /&gt;
Contact the Admin Team to access previously submitted slides.&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Completion Certificate ===&lt;br /&gt;
A certificate of completion and a letter of recommendation will be provided to all participants who successfully complete the program.&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
=== Contact ===&lt;br /&gt;
mazumder_lab@gwu.edu.&lt;br /&gt;
&amp;lt;hr&amp;gt;&lt;br /&gt;
=== Volunteers ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
|-&lt;br /&gt;
! Name&lt;br /&gt;
!Project&lt;br /&gt;
!Projects Interested&lt;br /&gt;
|-&lt;br /&gt;
| [https://www.linkedin.com/in/gracesjchong/ Grace Chong]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
# PredictMod&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/alma-ogunsina-4959072b1/ Alma Ogunsina]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
# ARGOS&lt;br /&gt;
# PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/diya-kamalabharathy-62557935a/ Diya Kamalabharathy]&lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# PredictMod Machine Learning&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/harivinay-prasad-reddy-gujjula-a06ba71bb/ Harivinay P. Gujjula]&lt;br /&gt;
|GlyGen curation&lt;br /&gt;
|&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# BioMarkerKB Biocuration&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/miao-wang-88b602290/Miao&amp;amp;#x20;Wang Miao Wang]&lt;br /&gt;
|ARGOS&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration Project Ideas&lt;br /&gt;
# FDA-ARGOS Computation and Pathogen Curation Project&lt;br /&gt;
# PredictMod Machine Learning Project Ideas&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/nahom-gebreselassie-1545ab336/ Nahom Abel]&lt;br /&gt;
|GlyGen curation&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# PredictMod&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/kajal-patel-cs/ Kajal Sanjaykumar Patel]&lt;br /&gt;
|GlyGen and PubMed project&lt;br /&gt;
|&lt;br /&gt;
#PredictMod&lt;br /&gt;
#BiomarkerKB&lt;br /&gt;
#GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/john-mccaffrey-b8850930a/ John McCaffrey]&lt;br /&gt;
|Biomarker curation&lt;br /&gt;
|&lt;br /&gt;
# PredictMod&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
# GlyGen Biocuration &lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/nathan-ressom/ Nathan Ressom]&lt;br /&gt;
|ARGOS&lt;br /&gt;
|&lt;br /&gt;
# PredictMod &lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/aaron-ressom/ Aaron Ressom] &lt;br /&gt;
|PredictMod&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB &lt;br /&gt;
# PredictMod &lt;br /&gt;
# GlyGen&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/akale-kinfe/ Akale Kinfe]&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|Aise Arpinar &lt;br /&gt;
|GlyGen curation&lt;br /&gt;
|&lt;br /&gt;
# GlyGen Biocuration&lt;br /&gt;
# BiomarkerKB Biocuration&lt;br /&gt;
# GlyGen Publication Analysis&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/piyush-pandey-906b582b5/ Piyush Pandey]&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB Biocuration &lt;br /&gt;
# PredictMod &lt;br /&gt;
# GlyGen Biocuration &lt;br /&gt;
|-&lt;br /&gt;
|[http://www.linkedin.com/in/filmawit-zeru-203272363 Filmawit Zeru]&lt;br /&gt;
|GlycoSiteMiner project&lt;br /&gt;
|&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
# GlyGen&lt;br /&gt;
# ARGOS&lt;br /&gt;
|-&lt;br /&gt;
|[https://www.linkedin.com/in/mathias-belay-03b51a2a3/ Mathias Belay]&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
# GlyGen&lt;br /&gt;
# PredictMod&lt;br /&gt;
# BiomarkerKB&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=813</id>
		<title>ARGOSQC Usage Tutorial</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=813"/>
		<updated>2025-04-25T20:11:20Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Where to locate NCBI Information for the inputs: */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
HIVE3 one-click pipeline tutorial for the FDA HIVE instance. This protocol will guide the user in running single and batch-mode QC computations. HIVE3 is an instance of HIVE not owned by the FDA which can be directly modified by Vahan or others on our team with permissions.&lt;br /&gt;
&lt;br /&gt;
= Required User Information =&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Protocol Version&#039;&#039;&#039; &lt;br /&gt;
|1.0&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Instance&#039;&#039;&#039;&lt;br /&gt;
|3&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Link&#039;&#039;&#039;&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=login&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Overview =&lt;br /&gt;
We constructed a QC one-click pipeline that takes user specified organism information and combines the 3 core ARGOS workflows to produce 5 different result datasets in JSON format (Figure 1). 3 out of the 5 result JSONs (assemblyQC, ngsQC, and biosampleMeta) have been updated and added to data.argosdb.&lt;br /&gt;
&lt;br /&gt;
To register your account, navigate to the link under the “Required User Information”. At the top right there will be a tab saying “register”. Fill out the appropriate fields and submit. Once submitted, please email mazumder_lab@gwu.edu so we can verify your account.&lt;br /&gt;
&lt;br /&gt;
The ARGOS pipeline can be accessed via the dropdown menu ‘Projects’ at the upper right hand screen and then under ‘Argos’. The pipeline in HIVE3 is located in the “Required User Information” section in the beginning of this protocol. &lt;br /&gt;
&lt;br /&gt;
After a successful login, you will be navigated to the home page. Use the menu at the top right corner under projects to access the ARGOS pipeline or use this URL to access the ARGOS QC pipeline on HIVE3:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=argos-alqc&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[File:Screenshot figure 1.png|thumb|667x667px|&#039;&#039;&#039;Figure 1.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline. |none]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Input values:&#039;&#039; ==&lt;br /&gt;
On the ARGOS Pipeline input setting page, under the General tab, is where our data inputs for a single and batch computation will be taken. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Name:&#039;&#039;&#039; Give the computation a name. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Folder:&#039;&#039;&#039; Give the folder where your computations, data, and steps will be stored.&lt;br /&gt;
&lt;br /&gt;
* Can use: _ or -  or &amp;amp;  and letters + numbers&lt;br /&gt;
* Cannot use: / : ; , \ “ ” ‘ ’ &lt;br /&gt;
* Yet, you use / to create a sub folder, but that is not recommended. Manually moving the subfolder is best.&lt;br /&gt;
* Ex folder: Influenza A (h5n1)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reads:&#039;&#039;&#039; information needed for ngsQC &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;SRR&#039;&#039;&#039;: the SRR accession number, can be multiple per organism by using a “,” or populate extra fields by clicking on the gray + sign. This tool uses the NCBI SRA Fasterq function to grab the fastq files directly from NCBI without the user needing to import them to HIVE.  &lt;br /&gt;
* &#039;&#039;&#039;HIVE reads:&#039;&#039;&#039; Drop down menu can select reads already uploaded into HIVE, either from previous computations or manual uploads. &lt;br /&gt;
* See in Figure 2a&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the SRR input box, rather the HIVE IDs menu, but you can if you want to, it will just search within HIVE. See ngsQC Protocol for how to upload SRR information using external downloader. This external downloader process is the same as in HIVE2 and 1.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reference:&#039;&#039;&#039; Information used for the assemblyQC portion of the algorithm.&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Reference accession:&#039;&#039;&#039; This is the REFSEQ or Genbank accession number from NCBI or Genbank.&lt;br /&gt;
* &#039;&#039;&#039;Assembly ID&#039;&#039;&#039;: This is the ASSEMBLY accession number from NCBI. &lt;br /&gt;
* &#039;&#039;&#039;HIVE genome:&#039;&#039;&#039; Use the drop down menu to select a reference genome that has already been uploaded into HIVE. &lt;br /&gt;
* See in Figure 2b&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the assembly ID or reference accession input box, but you can if you want to. See AssemblyQC Protocol for how to upload assembly information using an external downloader or local upload. This process is the same as what is done in HIVE2.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Metadata: &#039;&#039;&#039; Used to grab information necessary to fill out the BiosampleMeta_HIVE document.&lt;br /&gt;
&lt;br /&gt;
* Biosample Accession: The optional accession number for the Biosample that was reported to be used when creating the assembly and will be linked to the SRR fastq files used for the ngsQC portion of the algorithm. This step is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Coding Table:&#039;&#039;&#039; Dropdown of genetic codon tables to be used for your computation, depending on the organism to be computed. The default is human, viral (Standard).&lt;br /&gt;
&lt;br /&gt;
* Tip: NCBI Taxonomy will list the codon table for each organism on their taxonomy page.&lt;br /&gt;
[[File:Screenshot figure 2 a.png|none|thumb|657x657px|&#039;&#039;&#039;Figure 2.&#039;&#039;&#039; The ARGOS_QC algorithm page input with all NCBI information for a test organism, Salmonella enterica.]]&lt;br /&gt;
[[File:Screenshot fig 2a.png|none|thumb|508x508px|&#039;&#039;&#039;Figure 2 a)&#039;&#039;&#039; The SRR accession field contains both SRR fastqs for the organism that correspond to the biosample.]]&lt;br /&gt;
[[File:Screenshot figure 2b.png|none|thumb|502x502px|&#039;&#039;&#039;Figure 2 b)&#039;&#039;&#039; The &#039;&#039;&#039;Reference Accession&#039;&#039;&#039; is the RefSeq Nucleotide accession number from NCBI, the &#039;&#039;&#039;Assembly ID&#039;&#039;&#039; is the assembly accession number from NCBI, and the Biosample Accession is the Biosample accession number from NCBI.                                         &#039;&#039;&#039;a &amp;amp; b)&#039;&#039;&#039; &#039;&#039;&#039;HIVE IDs&#039;&#039;&#039; are accessions that are already uploaded into HIVE and the algorithm will automatically select this information as opposed to pulling information from outside sources. ]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Where to locate NCBI Information for the inputs:&#039;&#039; ==&lt;br /&gt;
Navigate to the NCBI legacy assembly page for your organism. Here you can find all of the information to be used for your computations.&lt;br /&gt;
[[File:Screenshot fig 3.png|none|thumb|518x518px|&#039;&#039;&#039;Figure 3.&#039;&#039;&#039; The information shown in the NCBI legacy assembly page. The RefSeq assembly accession corresponds to our organism of interest and will be used to fill out the &#039;&#039;&#039;“Assembly ID”&#039;&#039;&#039; input field on the HIVE3 ARGOS_QC input page. Please note that the bioproject matches the accession for the FDA_ARGOS bioproject, and there are 2 sequencing technologies listed, meaning that there will most likely be two SRR submissions that we can find on the SRA page (see Figure 5 and 6). ]]&lt;br /&gt;
[[File:Screenshot fig 4.png|none|thumb|670x670px|&#039;&#039;&#039;Figure 4&#039;&#039;&#039;. The bottom section of the legacy assembly page for our test organism. The column labeled RefSeq sequence lists the DNA RefSeq for our test organism which we will use for the “&#039;&#039;&#039;Reference Accession&#039;&#039;&#039;” field on the HIVE3 ARGOS_QC input page. ]]&lt;br /&gt;
[[File:Screenshot fig 5.png|none|thumb|625x625px|&#039;&#039;&#039;Figure 5.&#039;&#039;&#039; Clicking on the Biosample accession number seen in Figure 3 will redirect you to the NCBI Biosample page for this organism. Under &#039;&#039;&#039;“Related Information”&#039;&#039;&#039;, click &#039;&#039;&#039;“SRA”&#039;&#039;&#039; to navigate to the NCBI SRA page for this biosample.]]&lt;br /&gt;
[[File:Screenshot fig 6.png|none|thumb|634x634px|&#039;&#039;&#039;Figure 6.&#039;&#039;&#039; The SRA page lists different sequencing links. Each link is reported to be sequenced on different platforms; either for Illumina or for the PacBio platform. This is common. The methodology behind using the different platforms is to gather insight for the assembly at different perspectives and levels. Illumina sequences DNA as multiple short reads that can be used to create an accurate reconstruction of the genomic sections analyzed by estimating the average/best fit nucleotide sequence. PacBio is a long read sequencer that takes “movies” of the DNA sequence as it moves through the technology and captures the sequence in one go from start to finish. The long read sequence then acts as a map for the short accurate short read sequences that need to be assembled. Therefore, it is important to use all of the links reported in our QC pipeline.]]&lt;br /&gt;
[[File:Screenshot fig 7.png|none|thumb|635x635px|&#039;&#039;&#039;Figure 7.&#039;&#039;&#039; Clicking on the bottom link under the &#039;&#039;&#039;Runs&#039;&#039;&#039; section from the SRA page shown in &#039;&#039;&#039;Figure 6&#039;&#039;&#039; will redirect to the page containing the SRR file. Copy and paste the SRR accession number into the input field in the HIVE3 pipeline labeled “SRR”. You will need to do this for both (or more) SRR accessions. Check that the bioproject, biosample, and organism name all coincide with our test organism.]]&lt;br /&gt;
&lt;br /&gt;
= Single QC Computation =&lt;br /&gt;
A single QC computation will allow for assemblyQC, biosampleQC, and ngsQC to be performed on one organism with one assembly, but can include multiple SRR ids. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Input the name of the computation and the name of the folder you wish to store the computation in. If you would like to add to a pre-existing folder, input the exact folder name in the folder input field. It is case sensitive. See input criteria at the beginning of this protocol.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; Under the dropdown menu for Reads, select which input you will use. &lt;br /&gt;
&lt;br /&gt;
* For SRRs, type or paste in the SRR id. If there is more than one SRR id, click on the gray + sign to populate a new input field or use a , .&lt;br /&gt;
** Troubleshooting: if the computation fails, try removing the spaces in between the commas and SRR ids. No spaces.&lt;br /&gt;
* For HIVE IDs, click on the HIVE ID option from the dropdown menu. Click on the gray dropdown menu arrow next to HIVE reads. A pop-up window will open, as seen in Figure 9. &lt;br /&gt;
** Click on the ids that you wish to use in the computation. Use Ctrl + shift to highlight multiple ids.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 8.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline filled out for a single QC computation of our test organism, Salmonella enterica. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 9.&#039;&#039;&#039; Pop-up window for SRR HIVE ID selection. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Next to Reference, select the input object you would like to use for the computation.&lt;br /&gt;
&lt;br /&gt;
* For Reference Accession, type or paste in the id you wish to use.&lt;br /&gt;
* For Assembly IDs, type or paste in the id you wish to use. &lt;br /&gt;
* For HIVE Genomes, refer to Step 2 above on how to select a HIVE id. It is the same process.&lt;br /&gt;
* Refer to the beginning of this protocol for what ids can be inputted.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Next under BioSample Accessions, paste in the biosample ID you would like to use for the computation. This is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Lastly, select from the Coding Table dropdown menu the genetic code you would like to use for your computation, if applicable. The default is “human, viral (Standard)”.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the information has been inputted correctly, click on the big blue Submit button in the middle, as seen in Figure 8.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 7:&#039;&#039;&#039; Once your computation has been submitted, you can click on the Home tab found in the top left to go back to the homepage and see the workflow.&lt;br /&gt;
&lt;br /&gt;
= Batch Mode Computation =&lt;br /&gt;
Batch mode operates by a user-specified ratio of groups. With the help of semicolons and commas, the ratio would be 1:1:1 for a batch mode computation. It is 1:1:1, because we are clustering them by ; so the pipeline recognizes the ids between the ; as one computation. It would be one cluster of SRRs to one assembly to one biosample, that is one computation. There is a colorful and highlighted example below displaying the syntax for the inputs.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;They would be grouped for computations like this example:&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
To separate between batches, use a semicolon “;” in between the IDs. A comma denotes separate IDs but semicolons as separate batches. These will be inputted in the General tab of the pipeline, same as single computation. Within each field, this is how the above example would look like in batch mode:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456, SRR0123457, SRR0123458, SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039; SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
Assembly IDs: GCA_0011223344.1&#039;&#039;&#039;;&#039;&#039;&#039; GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
BioSample Accessions: SAMN110654321, SAMN110654322&#039;&#039;&#039;;&#039;&#039;&#039; SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Notice the semicolon separation according to the example above. The commas separate the ids, and the semicolon the batch.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Troubleshooting Note:&#039;&#039;&#039; If your computation fails or if there is an error, remove the spaces between the , and ; .Previously, this has thrown an error but has been fixed, but worth a shot if your computation fails. It would look like:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456,SRR0123457,SRR0123458,SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039;SRR0123451,SRR0123452,SRR0123453,SRR0123454&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Navigate to the tab title Batch, Figure 10. This can be found on the ARGOS input settings page, Figure 1., next to the General tab. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; For the parameter “batch service&amp;quot; at the bottom select, from the dropdown menu batch mode. This will have the pipeline set to Batch Mode rather than single computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 10.&#039;&#039;&#039; Batch mode input settings window.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Selecting the parameters. Click on the drop down menu next to the text “Parameter list”.&lt;br /&gt;
&lt;br /&gt;
* Use the black plus button next to ‘Parameter List’ to populate an entry field.&lt;br /&gt;
* Select from the dropdown field the correct parameter based on the input field you used in the general input page. This can be seen in Figure 11.&lt;br /&gt;
* For example, if you pasted in SRR Ids you would choose the parameter SRR IDs. If you chose HIVEIDs you would select HIVE IDs from the dropdown.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 11.&#039;&#039;&#039; Dropdown menu from the Batch tab displaying the parameter options. Select the batch parameter option based on what the input information is in the general tab. For example, if you entered SRR ids, select SRR IDs from the dropdown. If you input a reference ID, select Reference IDs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Input the ratio for the batch service. &lt;br /&gt;
&lt;br /&gt;
* For computations in batch mode in the one-click pipeline, the computations are separated by semicolon “;” and the IDs within the computations by a comma “,”. Since the workflow will parse the computations and recognize the IDs between the “;” as one computation, the ratio will be 1:1:1. &lt;br /&gt;
* If the ratio is 1:1:1 then enter the value 1 for each box.&lt;br /&gt;
* One set of SRRs to one assembly to one biosample. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Inputting the information correctly in each field. Navigate back to the input settings page, Figure 1. The same page that you had used for the single computations will be used, but the only difference will be the semicolons and commas. The example below will visually show you how the information will be inputted for a batch mode.&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Batch Mode Parameter breakdown:&#039;&#039; ==&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Again, this is very similar to single computations, except that the batch mode will use semicolons and commas to separate the ids.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the input information is complete, hit the blue button Submit. You may exit the Argos pipeline window by hitting “Home” on the top left corner.&lt;br /&gt;
&lt;br /&gt;
= QC Computation Results =&lt;br /&gt;
Once you have submitted your computations, either single or batch, you should see the pipeline workflow in your Inbox or All Objects, as shown in Figure 12. You can also view the pipeline by clicking on the “workflows” tab also seen in Figure 12.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 12.&#039;&#039;&#039;  The pipeline workflow displayed in the user’s inbox.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
As the workflow progresses, your computations will be stored in the folder that you named from the beginning of this protocol. To view the contents of the folder, simply click on the plus sign next to the folder or the folder name to open.&lt;br /&gt;
&lt;br /&gt;
Once your computations are complete the QC outputs will be stored in JSON file format from the computation “&#039;&#039;&#039;Post-Alignment Quality Controls&#039;&#039;&#039;” or under the &#039;&#039;&#039;“CFlow”&#039;&#039;&#039; workflow. P-A QC can be found in the folder you specified for the computation or CFlow in All Objects. To view the JSONs click on the name so that it is highlighted blue and click on the tab from the bottom menu named “Available Downloads”.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 13.&#039;&#039;&#039; The available downloads tab and the 5 JSON files that are the QC outputs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
There will be 5 files reported in JSON format. Click the blue/green download icon next to each file to see the results. The file labeled &#039;&#039;&#039;qcAll.json&#039;&#039;&#039; will have our assemblyQC results. &#039;&#039;&#039;qcNGS.json&#039;&#039;&#039; will have our ngsQC results and &#039;&#039;&#039;biosample.json&#039;&#039;&#039; the biosample information. We currently do not submit qcPos.json or refAnnot.json to the ARGOS DB, but the information is there to better help you understand your computation.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_7.png&amp;diff=812</id>
		<title>File:Screenshot fig 7.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_7.png&amp;diff=812"/>
		<updated>2025-04-25T20:10:49Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;fig&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_6.png&amp;diff=811</id>
		<title>File:Screenshot fig 6.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_6.png&amp;diff=811"/>
		<updated>2025-04-25T20:09:25Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;fig&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_5.png&amp;diff=810</id>
		<title>File:Screenshot fig 5.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_5.png&amp;diff=810"/>
		<updated>2025-04-25T20:07:37Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;fig&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_4.png&amp;diff=809</id>
		<title>File:Screenshot fig 4.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_4.png&amp;diff=809"/>
		<updated>2025-04-25T20:05:39Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;fig&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=808</id>
		<title>ARGOSQC Usage Tutorial</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=808"/>
		<updated>2025-04-25T20:03:53Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Where to locate NCBI Information for the inputs: */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
HIVE3 one-click pipeline tutorial for the FDA HIVE instance. This protocol will guide the user in running single and batch-mode QC computations. HIVE3 is an instance of HIVE not owned by the FDA which can be directly modified by Vahan or others on our team with permissions.&lt;br /&gt;
&lt;br /&gt;
= Required User Information =&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Protocol Version&#039;&#039;&#039; &lt;br /&gt;
|1.0&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Instance&#039;&#039;&#039;&lt;br /&gt;
|3&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Link&#039;&#039;&#039;&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=login&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Overview =&lt;br /&gt;
We constructed a QC one-click pipeline that takes user specified organism information and combines the 3 core ARGOS workflows to produce 5 different result datasets in JSON format (Figure 1). 3 out of the 5 result JSONs (assemblyQC, ngsQC, and biosampleMeta) have been updated and added to data.argosdb.&lt;br /&gt;
&lt;br /&gt;
To register your account, navigate to the link under the “Required User Information”. At the top right there will be a tab saying “register”. Fill out the appropriate fields and submit. Once submitted, please email mazumder_lab@gwu.edu so we can verify your account.&lt;br /&gt;
&lt;br /&gt;
The ARGOS pipeline can be accessed via the dropdown menu ‘Projects’ at the upper right hand screen and then under ‘Argos’. The pipeline in HIVE3 is located in the “Required User Information” section in the beginning of this protocol. &lt;br /&gt;
&lt;br /&gt;
After a successful login, you will be navigated to the home page. Use the menu at the top right corner under projects to access the ARGOS pipeline or use this URL to access the ARGOS QC pipeline on HIVE3:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=argos-alqc&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[File:Screenshot figure 1.png|thumb|667x667px|&#039;&#039;&#039;Figure 1.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline. |none]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Input values:&#039;&#039; ==&lt;br /&gt;
On the ARGOS Pipeline input setting page, under the General tab, is where our data inputs for a single and batch computation will be taken. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Name:&#039;&#039;&#039; Give the computation a name. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Folder:&#039;&#039;&#039; Give the folder where your computations, data, and steps will be stored.&lt;br /&gt;
&lt;br /&gt;
* Can use: _ or -  or &amp;amp;  and letters + numbers&lt;br /&gt;
* Cannot use: / : ; , \ “ ” ‘ ’ &lt;br /&gt;
* Yet, you use / to create a sub folder, but that is not recommended. Manually moving the subfolder is best.&lt;br /&gt;
* Ex folder: Influenza A (h5n1)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reads:&#039;&#039;&#039; information needed for ngsQC &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;SRR&#039;&#039;&#039;: the SRR accession number, can be multiple per organism by using a “,” or populate extra fields by clicking on the gray + sign. This tool uses the NCBI SRA Fasterq function to grab the fastq files directly from NCBI without the user needing to import them to HIVE.  &lt;br /&gt;
* &#039;&#039;&#039;HIVE reads:&#039;&#039;&#039; Drop down menu can select reads already uploaded into HIVE, either from previous computations or manual uploads. &lt;br /&gt;
* See in Figure 2a&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the SRR input box, rather the HIVE IDs menu, but you can if you want to, it will just search within HIVE. See ngsQC Protocol for how to upload SRR information using external downloader. This external downloader process is the same as in HIVE2 and 1.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reference:&#039;&#039;&#039; Information used for the assemblyQC portion of the algorithm.&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Reference accession:&#039;&#039;&#039; This is the REFSEQ or Genbank accession number from NCBI or Genbank.&lt;br /&gt;
* &#039;&#039;&#039;Assembly ID&#039;&#039;&#039;: This is the ASSEMBLY accession number from NCBI. &lt;br /&gt;
* &#039;&#039;&#039;HIVE genome:&#039;&#039;&#039; Use the drop down menu to select a reference genome that has already been uploaded into HIVE. &lt;br /&gt;
* See in Figure 2b&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the assembly ID or reference accession input box, but you can if you want to. See AssemblyQC Protocol for how to upload assembly information using an external downloader or local upload. This process is the same as what is done in HIVE2.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Metadata: &#039;&#039;&#039; Used to grab information necessary to fill out the BiosampleMeta_HIVE document.&lt;br /&gt;
&lt;br /&gt;
* Biosample Accession: The optional accession number for the Biosample that was reported to be used when creating the assembly and will be linked to the SRR fastq files used for the ngsQC portion of the algorithm. This step is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Coding Table:&#039;&#039;&#039; Dropdown of genetic codon tables to be used for your computation, depending on the organism to be computed. The default is human, viral (Standard).&lt;br /&gt;
&lt;br /&gt;
* Tip: NCBI Taxonomy will list the codon table for each organism on their taxonomy page.&lt;br /&gt;
[[File:Screenshot figure 2 a.png|none|thumb|657x657px|&#039;&#039;&#039;Figure 2.&#039;&#039;&#039; The ARGOS_QC algorithm page input with all NCBI information for a test organism, Salmonella enterica.]]&lt;br /&gt;
[[File:Screenshot fig 2a.png|none|thumb|508x508px|&#039;&#039;&#039;Figure 2 a)&#039;&#039;&#039; The SRR accession field contains both SRR fastqs for the organism that correspond to the biosample.]]&lt;br /&gt;
[[File:Screenshot figure 2b.png|none|thumb|502x502px|&#039;&#039;&#039;Figure 2 b)&#039;&#039;&#039; The &#039;&#039;&#039;Reference Accession&#039;&#039;&#039; is the RefSeq Nucleotide accession number from NCBI, the &#039;&#039;&#039;Assembly ID&#039;&#039;&#039; is the assembly accession number from NCBI, and the Biosample Accession is the Biosample accession number from NCBI.                                         &#039;&#039;&#039;a &amp;amp; b)&#039;&#039;&#039; &#039;&#039;&#039;HIVE IDs&#039;&#039;&#039; are accessions that are already uploaded into HIVE and the algorithm will automatically select this information as opposed to pulling information from outside sources. ]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Where to locate NCBI Information for the inputs:&#039;&#039; ==&lt;br /&gt;
Navigate to the NCBI legacy assembly page for your organism. Here you can find all of the information to be used for your computations.&lt;br /&gt;
[[File:Screenshot fig 3.png|none|thumb|518x518px|&#039;&#039;&#039;Figure 3.&#039;&#039;&#039; The information shown in the NCBI legacy assembly page. The RefSeq assembly accession corresponds to our organism of interest and will be used to fill out the &#039;&#039;&#039;“Assembly ID”&#039;&#039;&#039; input field on the HIVE3 ARGOS_QC input page. Please note that the bioproject matches the accession for the FDA_ARGOS bioproject, and there are 2 sequencing technologies listed, meaning that there will most likely be two SRR submissions that we can find on the SRA page (see Figure 5 and 6). ]]&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 4&#039;&#039;&#039;. The bottom section of the legacy assembly page for our test organism. The column labeled RefSeq sequence lists the DNA RefSeq for our test organism which we will use for the “&#039;&#039;&#039;Reference Accession&#039;&#039;&#039;” field on the HIVE3 ARGOS_QC input page. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 5.&#039;&#039;&#039; Clicking on the Biosample accession number seen in Figure 3 will redirect you to the NCBI Biosample page for this organism. Under &#039;&#039;&#039;“Related Information”&#039;&#039;&#039;, click &#039;&#039;&#039;“SRA”&#039;&#039;&#039; to navigate to the NCBI SRA page for this biosample.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 6.&#039;&#039;&#039; The SRA page lists different sequencing links. Each link is reported to be sequenced on different platforms; either for Illumina or for the PacBio platform. This is common. The methodology behind using the different platforms is to gather insight for the assembly at different perspectives and levels. Illumina sequences DNA as multiple short reads that can be used to create an accurate reconstruction of the genomic sections analyzed by estimating the average/best fit nucleotide sequence. PacBio is a long read sequencer that takes “movies” of the DNA sequence as it moves through the technology and captures the sequence in one go from start to finish. The long read sequence then acts as a map for the short accurate short read sequences that need to be assembled. Therefore, it is important to use all of the links reported in our QC pipeline.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 7.&#039;&#039;&#039; Clicking on the bottom link under the &#039;&#039;&#039;Runs&#039;&#039;&#039; section from the SRA page shown in &#039;&#039;&#039;Figure 6&#039;&#039;&#039; will redirect to the page containing the SRR file. Copy and paste the SRR accession number into the input field in the HIVE3 pipeline labeled “SRR”. You will need to do this for both (or more) SRR accessions. Check that the bioproject, biosample, and organism name all coincide with our test organism.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Single QC Computation =&lt;br /&gt;
A single QC computation will allow for assemblyQC, biosampleQC, and ngsQC to be performed on one organism with one assembly, but can include multiple SRR ids. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Input the name of the computation and the name of the folder you wish to store the computation in. If you would like to add to a pre-existing folder, input the exact folder name in the folder input field. It is case sensitive. See input criteria at the beginning of this protocol.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; Under the dropdown menu for Reads, select which input you will use. &lt;br /&gt;
&lt;br /&gt;
* For SRRs, type or paste in the SRR id. If there is more than one SRR id, click on the gray + sign to populate a new input field or use a , .&lt;br /&gt;
** Troubleshooting: if the computation fails, try removing the spaces in between the commas and SRR ids. No spaces.&lt;br /&gt;
* For HIVE IDs, click on the HIVE ID option from the dropdown menu. Click on the gray dropdown menu arrow next to HIVE reads. A pop-up window will open, as seen in Figure 9. &lt;br /&gt;
** Click on the ids that you wish to use in the computation. Use Ctrl + shift to highlight multiple ids.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 8.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline filled out for a single QC computation of our test organism, Salmonella enterica. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 9.&#039;&#039;&#039; Pop-up window for SRR HIVE ID selection. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Next to Reference, select the input object you would like to use for the computation.&lt;br /&gt;
&lt;br /&gt;
* For Reference Accession, type or paste in the id you wish to use.&lt;br /&gt;
* For Assembly IDs, type or paste in the id you wish to use. &lt;br /&gt;
* For HIVE Genomes, refer to Step 2 above on how to select a HIVE id. It is the same process.&lt;br /&gt;
* Refer to the beginning of this protocol for what ids can be inputted.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Next under BioSample Accessions, paste in the biosample ID you would like to use for the computation. This is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Lastly, select from the Coding Table dropdown menu the genetic code you would like to use for your computation, if applicable. The default is “human, viral (Standard)”.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the information has been inputted correctly, click on the big blue Submit button in the middle, as seen in Figure 8.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 7:&#039;&#039;&#039; Once your computation has been submitted, you can click on the Home tab found in the top left to go back to the homepage and see the workflow.&lt;br /&gt;
&lt;br /&gt;
= Batch Mode Computation =&lt;br /&gt;
Batch mode operates by a user-specified ratio of groups. With the help of semicolons and commas, the ratio would be 1:1:1 for a batch mode computation. It is 1:1:1, because we are clustering them by ; so the pipeline recognizes the ids between the ; as one computation. It would be one cluster of SRRs to one assembly to one biosample, that is one computation. There is a colorful and highlighted example below displaying the syntax for the inputs.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;They would be grouped for computations like this example:&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
To separate between batches, use a semicolon “;” in between the IDs. A comma denotes separate IDs but semicolons as separate batches. These will be inputted in the General tab of the pipeline, same as single computation. Within each field, this is how the above example would look like in batch mode:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456, SRR0123457, SRR0123458, SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039; SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
Assembly IDs: GCA_0011223344.1&#039;&#039;&#039;;&#039;&#039;&#039; GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
BioSample Accessions: SAMN110654321, SAMN110654322&#039;&#039;&#039;;&#039;&#039;&#039; SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Notice the semicolon separation according to the example above. The commas separate the ids, and the semicolon the batch.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Troubleshooting Note:&#039;&#039;&#039; If your computation fails or if there is an error, remove the spaces between the , and ; .Previously, this has thrown an error but has been fixed, but worth a shot if your computation fails. It would look like:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456,SRR0123457,SRR0123458,SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039;SRR0123451,SRR0123452,SRR0123453,SRR0123454&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Navigate to the tab title Batch, Figure 10. This can be found on the ARGOS input settings page, Figure 1., next to the General tab. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; For the parameter “batch service&amp;quot; at the bottom select, from the dropdown menu batch mode. This will have the pipeline set to Batch Mode rather than single computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 10.&#039;&#039;&#039; Batch mode input settings window.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Selecting the parameters. Click on the drop down menu next to the text “Parameter list”.&lt;br /&gt;
&lt;br /&gt;
* Use the black plus button next to ‘Parameter List’ to populate an entry field.&lt;br /&gt;
* Select from the dropdown field the correct parameter based on the input field you used in the general input page. This can be seen in Figure 11.&lt;br /&gt;
* For example, if you pasted in SRR Ids you would choose the parameter SRR IDs. If you chose HIVEIDs you would select HIVE IDs from the dropdown.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 11.&#039;&#039;&#039; Dropdown menu from the Batch tab displaying the parameter options. Select the batch parameter option based on what the input information is in the general tab. For example, if you entered SRR ids, select SRR IDs from the dropdown. If you input a reference ID, select Reference IDs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Input the ratio for the batch service. &lt;br /&gt;
&lt;br /&gt;
* For computations in batch mode in the one-click pipeline, the computations are separated by semicolon “;” and the IDs within the computations by a comma “,”. Since the workflow will parse the computations and recognize the IDs between the “;” as one computation, the ratio will be 1:1:1. &lt;br /&gt;
* If the ratio is 1:1:1 then enter the value 1 for each box.&lt;br /&gt;
* One set of SRRs to one assembly to one biosample. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Inputting the information correctly in each field. Navigate back to the input settings page, Figure 1. The same page that you had used for the single computations will be used, but the only difference will be the semicolons and commas. The example below will visually show you how the information will be inputted for a batch mode.&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Batch Mode Parameter breakdown:&#039;&#039; ==&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Again, this is very similar to single computations, except that the batch mode will use semicolons and commas to separate the ids.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the input information is complete, hit the blue button Submit. You may exit the Argos pipeline window by hitting “Home” on the top left corner.&lt;br /&gt;
&lt;br /&gt;
= QC Computation Results =&lt;br /&gt;
Once you have submitted your computations, either single or batch, you should see the pipeline workflow in your Inbox or All Objects, as shown in Figure 12. You can also view the pipeline by clicking on the “workflows” tab also seen in Figure 12.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 12.&#039;&#039;&#039;  The pipeline workflow displayed in the user’s inbox.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
As the workflow progresses, your computations will be stored in the folder that you named from the beginning of this protocol. To view the contents of the folder, simply click on the plus sign next to the folder or the folder name to open.&lt;br /&gt;
&lt;br /&gt;
Once your computations are complete the QC outputs will be stored in JSON file format from the computation “&#039;&#039;&#039;Post-Alignment Quality Controls&#039;&#039;&#039;” or under the &#039;&#039;&#039;“CFlow”&#039;&#039;&#039; workflow. P-A QC can be found in the folder you specified for the computation or CFlow in All Objects. To view the JSONs click on the name so that it is highlighted blue and click on the tab from the bottom menu named “Available Downloads”.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 13.&#039;&#039;&#039; The available downloads tab and the 5 JSON files that are the QC outputs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
There will be 5 files reported in JSON format. Click the blue/green download icon next to each file to see the results. The file labeled &#039;&#039;&#039;qcAll.json&#039;&#039;&#039; will have our assemblyQC results. &#039;&#039;&#039;qcNGS.json&#039;&#039;&#039; will have our ngsQC results and &#039;&#039;&#039;biosample.json&#039;&#039;&#039; the biosample information. We currently do not submit qcPos.json or refAnnot.json to the ARGOS DB, but the information is there to better help you understand your computation.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_3.png&amp;diff=807</id>
		<title>File:Screenshot fig 3.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_3.png&amp;diff=807"/>
		<updated>2025-04-25T20:03:06Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;fig&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=806</id>
		<title>ARGOSQC Usage Tutorial</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=806"/>
		<updated>2025-04-25T19:58:31Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Input values: */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
HIVE3 one-click pipeline tutorial for the FDA HIVE instance. This protocol will guide the user in running single and batch-mode QC computations. HIVE3 is an instance of HIVE not owned by the FDA which can be directly modified by Vahan or others on our team with permissions.&lt;br /&gt;
&lt;br /&gt;
= Required User Information =&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Protocol Version&#039;&#039;&#039; &lt;br /&gt;
|1.0&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Instance&#039;&#039;&#039;&lt;br /&gt;
|3&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Link&#039;&#039;&#039;&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=login&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Overview =&lt;br /&gt;
We constructed a QC one-click pipeline that takes user specified organism information and combines the 3 core ARGOS workflows to produce 5 different result datasets in JSON format (Figure 1). 3 out of the 5 result JSONs (assemblyQC, ngsQC, and biosampleMeta) have been updated and added to data.argosdb.&lt;br /&gt;
&lt;br /&gt;
To register your account, navigate to the link under the “Required User Information”. At the top right there will be a tab saying “register”. Fill out the appropriate fields and submit. Once submitted, please email mazumder_lab@gwu.edu so we can verify your account.&lt;br /&gt;
&lt;br /&gt;
The ARGOS pipeline can be accessed via the dropdown menu ‘Projects’ at the upper right hand screen and then under ‘Argos’. The pipeline in HIVE3 is located in the “Required User Information” section in the beginning of this protocol. &lt;br /&gt;
&lt;br /&gt;
After a successful login, you will be navigated to the home page. Use the menu at the top right corner under projects to access the ARGOS pipeline or use this URL to access the ARGOS QC pipeline on HIVE3:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=argos-alqc&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[File:Screenshot figure 1.png|thumb|667x667px|&#039;&#039;&#039;Figure 1.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline. |none]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Input values:&#039;&#039; ==&lt;br /&gt;
On the ARGOS Pipeline input setting page, under the General tab, is where our data inputs for a single and batch computation will be taken. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Name:&#039;&#039;&#039; Give the computation a name. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Folder:&#039;&#039;&#039; Give the folder where your computations, data, and steps will be stored.&lt;br /&gt;
&lt;br /&gt;
* Can use: _ or -  or &amp;amp;  and letters + numbers&lt;br /&gt;
* Cannot use: / : ; , \ “ ” ‘ ’ &lt;br /&gt;
* Yet, you use / to create a sub folder, but that is not recommended. Manually moving the subfolder is best.&lt;br /&gt;
* Ex folder: Influenza A (h5n1)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reads:&#039;&#039;&#039; information needed for ngsQC &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;SRR&#039;&#039;&#039;: the SRR accession number, can be multiple per organism by using a “,” or populate extra fields by clicking on the gray + sign. This tool uses the NCBI SRA Fasterq function to grab the fastq files directly from NCBI without the user needing to import them to HIVE.  &lt;br /&gt;
* &#039;&#039;&#039;HIVE reads:&#039;&#039;&#039; Drop down menu can select reads already uploaded into HIVE, either from previous computations or manual uploads. &lt;br /&gt;
* See in Figure 2a&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the SRR input box, rather the HIVE IDs menu, but you can if you want to, it will just search within HIVE. See ngsQC Protocol for how to upload SRR information using external downloader. This external downloader process is the same as in HIVE2 and 1.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reference:&#039;&#039;&#039; Information used for the assemblyQC portion of the algorithm.&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Reference accession:&#039;&#039;&#039; This is the REFSEQ or Genbank accession number from NCBI or Genbank.&lt;br /&gt;
* &#039;&#039;&#039;Assembly ID&#039;&#039;&#039;: This is the ASSEMBLY accession number from NCBI. &lt;br /&gt;
* &#039;&#039;&#039;HIVE genome:&#039;&#039;&#039; Use the drop down menu to select a reference genome that has already been uploaded into HIVE. &lt;br /&gt;
* See in Figure 2b&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the assembly ID or reference accession input box, but you can if you want to. See AssemblyQC Protocol for how to upload assembly information using an external downloader or local upload. This process is the same as what is done in HIVE2.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Metadata: &#039;&#039;&#039; Used to grab information necessary to fill out the BiosampleMeta_HIVE document.&lt;br /&gt;
&lt;br /&gt;
* Biosample Accession: The optional accession number for the Biosample that was reported to be used when creating the assembly and will be linked to the SRR fastq files used for the ngsQC portion of the algorithm. This step is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Coding Table:&#039;&#039;&#039; Dropdown of genetic codon tables to be used for your computation, depending on the organism to be computed. The default is human, viral (Standard).&lt;br /&gt;
&lt;br /&gt;
* Tip: NCBI Taxonomy will list the codon table for each organism on their taxonomy page.&lt;br /&gt;
[[File:Screenshot figure 2 a.png|none|thumb|657x657px|&#039;&#039;&#039;Figure 2.&#039;&#039;&#039; The ARGOS_QC algorithm page input with all NCBI information for a test organism, Salmonella enterica.]]&lt;br /&gt;
[[File:Screenshot fig 2a.png|none|thumb|508x508px|&#039;&#039;&#039;Figure 2 a)&#039;&#039;&#039; The SRR accession field contains both SRR fastqs for the organism that correspond to the biosample.]]&lt;br /&gt;
[[File:Screenshot figure 2b.png|none|thumb|502x502px|&#039;&#039;&#039;Figure 2 b)&#039;&#039;&#039; The &#039;&#039;&#039;Reference Accession&#039;&#039;&#039; is the RefSeq Nucleotide accession number from NCBI, the &#039;&#039;&#039;Assembly ID&#039;&#039;&#039; is the assembly accession number from NCBI, and the Biosample Accession is the Biosample accession number from NCBI.                                         &#039;&#039;&#039;a &amp;amp; b)&#039;&#039;&#039; &#039;&#039;&#039;HIVE IDs&#039;&#039;&#039; are accessions that are already uploaded into HIVE and the algorithm will automatically select this information as opposed to pulling information from outside sources. ]]&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Where to locate NCBI Information for the inputs:&#039;&#039; ==&lt;br /&gt;
Navigate to the NCBI legacy assembly page for your organism. Here you can find all of the information to be used for your computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 3.&#039;&#039;&#039; The information shown in the NCBI legacy assembly page. The RefSeq assembly accession corresponds to our organism of interest and will be used to fill out the &#039;&#039;&#039;“Assembly ID”&#039;&#039;&#039; input field on the HIVE3 ARGOS_QC input page. Please note that the bioproject matches the accession for the FDA_ARGOS bioproject, and there are 2 sequencing technologies listed, meaning that there will most likely be two SRR submissions that we can find on the SRA page (see Figure 5 and 6). &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 4&#039;&#039;&#039;. The bottom section of the legacy assembly page for our test organism. The column labeled RefSeq sequence lists the DNA RefSeq for our test organism which we will use for the “&#039;&#039;&#039;Reference Accession&#039;&#039;&#039;” field on the HIVE3 ARGOS_QC input page. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 5.&#039;&#039;&#039; Clicking on the Biosample accession number seen in Figure 3 will redirect you to the NCBI Biosample page for this organism. Under &#039;&#039;&#039;“Related Information”&#039;&#039;&#039;, click &#039;&#039;&#039;“SRA”&#039;&#039;&#039; to navigate to the NCBI SRA page for this biosample.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 6.&#039;&#039;&#039; The SRA page lists different sequencing links. Each link is reported to be sequenced on different platforms; either for Illumina or for the PacBio platform. This is common. The methodology behind using the different platforms is to gather insight for the assembly at different perspectives and levels. Illumina sequences DNA as multiple short reads that can be used to create an accurate reconstruction of the genomic sections analyzed by estimating the average/best fit nucleotide sequence. PacBio is a long read sequencer that takes “movies” of the DNA sequence as it moves through the technology and captures the sequence in one go from start to finish. The long read sequence then acts as a map for the short accurate short read sequences that need to be assembled. Therefore, it is important to use all of the links reported in our QC pipeline.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 7.&#039;&#039;&#039; Clicking on the bottom link under the &#039;&#039;&#039;Runs&#039;&#039;&#039; section from the SRA page shown in &#039;&#039;&#039;Figure 6&#039;&#039;&#039; will redirect to the page containing the SRR file. Copy and paste the SRR accession number into the input field in the HIVE3 pipeline labeled “SRR”. You will need to do this for both (or more) SRR accessions. Check that the bioproject, biosample, and organism name all coincide with our test organism.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Single QC Computation =&lt;br /&gt;
A single QC computation will allow for assemblyQC, biosampleQC, and ngsQC to be performed on one organism with one assembly, but can include multiple SRR ids. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Input the name of the computation and the name of the folder you wish to store the computation in. If you would like to add to a pre-existing folder, input the exact folder name in the folder input field. It is case sensitive. See input criteria at the beginning of this protocol.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; Under the dropdown menu for Reads, select which input you will use. &lt;br /&gt;
&lt;br /&gt;
* For SRRs, type or paste in the SRR id. If there is more than one SRR id, click on the gray + sign to populate a new input field or use a , .&lt;br /&gt;
** Troubleshooting: if the computation fails, try removing the spaces in between the commas and SRR ids. No spaces.&lt;br /&gt;
* For HIVE IDs, click on the HIVE ID option from the dropdown menu. Click on the gray dropdown menu arrow next to HIVE reads. A pop-up window will open, as seen in Figure 9. &lt;br /&gt;
** Click on the ids that you wish to use in the computation. Use Ctrl + shift to highlight multiple ids.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 8.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline filled out for a single QC computation of our test organism, Salmonella enterica. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 9.&#039;&#039;&#039; Pop-up window for SRR HIVE ID selection. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Next to Reference, select the input object you would like to use for the computation.&lt;br /&gt;
&lt;br /&gt;
* For Reference Accession, type or paste in the id you wish to use.&lt;br /&gt;
* For Assembly IDs, type or paste in the id you wish to use. &lt;br /&gt;
* For HIVE Genomes, refer to Step 2 above on how to select a HIVE id. It is the same process.&lt;br /&gt;
* Refer to the beginning of this protocol for what ids can be inputted.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Next under BioSample Accessions, paste in the biosample ID you would like to use for the computation. This is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Lastly, select from the Coding Table dropdown menu the genetic code you would like to use for your computation, if applicable. The default is “human, viral (Standard)”.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the information has been inputted correctly, click on the big blue Submit button in the middle, as seen in Figure 8.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 7:&#039;&#039;&#039; Once your computation has been submitted, you can click on the Home tab found in the top left to go back to the homepage and see the workflow.&lt;br /&gt;
&lt;br /&gt;
= Batch Mode Computation =&lt;br /&gt;
Batch mode operates by a user-specified ratio of groups. With the help of semicolons and commas, the ratio would be 1:1:1 for a batch mode computation. It is 1:1:1, because we are clustering them by ; so the pipeline recognizes the ids between the ; as one computation. It would be one cluster of SRRs to one assembly to one biosample, that is one computation. There is a colorful and highlighted example below displaying the syntax for the inputs.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;They would be grouped for computations like this example:&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
To separate between batches, use a semicolon “;” in between the IDs. A comma denotes separate IDs but semicolons as separate batches. These will be inputted in the General tab of the pipeline, same as single computation. Within each field, this is how the above example would look like in batch mode:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456, SRR0123457, SRR0123458, SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039; SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
Assembly IDs: GCA_0011223344.1&#039;&#039;&#039;;&#039;&#039;&#039; GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
BioSample Accessions: SAMN110654321, SAMN110654322&#039;&#039;&#039;;&#039;&#039;&#039; SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Notice the semicolon separation according to the example above. The commas separate the ids, and the semicolon the batch.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Troubleshooting Note:&#039;&#039;&#039; If your computation fails or if there is an error, remove the spaces between the , and ; .Previously, this has thrown an error but has been fixed, but worth a shot if your computation fails. It would look like:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456,SRR0123457,SRR0123458,SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039;SRR0123451,SRR0123452,SRR0123453,SRR0123454&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Navigate to the tab title Batch, Figure 10. This can be found on the ARGOS input settings page, Figure 1., next to the General tab. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; For the parameter “batch service&amp;quot; at the bottom select, from the dropdown menu batch mode. This will have the pipeline set to Batch Mode rather than single computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 10.&#039;&#039;&#039; Batch mode input settings window.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Selecting the parameters. Click on the drop down menu next to the text “Parameter list”.&lt;br /&gt;
&lt;br /&gt;
* Use the black plus button next to ‘Parameter List’ to populate an entry field.&lt;br /&gt;
* Select from the dropdown field the correct parameter based on the input field you used in the general input page. This can be seen in Figure 11.&lt;br /&gt;
* For example, if you pasted in SRR Ids you would choose the parameter SRR IDs. If you chose HIVEIDs you would select HIVE IDs from the dropdown.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 11.&#039;&#039;&#039; Dropdown menu from the Batch tab displaying the parameter options. Select the batch parameter option based on what the input information is in the general tab. For example, if you entered SRR ids, select SRR IDs from the dropdown. If you input a reference ID, select Reference IDs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Input the ratio for the batch service. &lt;br /&gt;
&lt;br /&gt;
* For computations in batch mode in the one-click pipeline, the computations are separated by semicolon “;” and the IDs within the computations by a comma “,”. Since the workflow will parse the computations and recognize the IDs between the “;” as one computation, the ratio will be 1:1:1. &lt;br /&gt;
* If the ratio is 1:1:1 then enter the value 1 for each box.&lt;br /&gt;
* One set of SRRs to one assembly to one biosample. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Inputting the information correctly in each field. Navigate back to the input settings page, Figure 1. The same page that you had used for the single computations will be used, but the only difference will be the semicolons and commas. The example below will visually show you how the information will be inputted for a batch mode.&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Batch Mode Parameter breakdown:&#039;&#039; ==&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Again, this is very similar to single computations, except that the batch mode will use semicolons and commas to separate the ids.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the input information is complete, hit the blue button Submit. You may exit the Argos pipeline window by hitting “Home” on the top left corner.&lt;br /&gt;
&lt;br /&gt;
= QC Computation Results =&lt;br /&gt;
Once you have submitted your computations, either single or batch, you should see the pipeline workflow in your Inbox or All Objects, as shown in Figure 12. You can also view the pipeline by clicking on the “workflows” tab also seen in Figure 12.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 12.&#039;&#039;&#039;  The pipeline workflow displayed in the user’s inbox.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
As the workflow progresses, your computations will be stored in the folder that you named from the beginning of this protocol. To view the contents of the folder, simply click on the plus sign next to the folder or the folder name to open.&lt;br /&gt;
&lt;br /&gt;
Once your computations are complete the QC outputs will be stored in JSON file format from the computation “&#039;&#039;&#039;Post-Alignment Quality Controls&#039;&#039;&#039;” or under the &#039;&#039;&#039;“CFlow”&#039;&#039;&#039; workflow. P-A QC can be found in the folder you specified for the computation or CFlow in All Objects. To view the JSONs click on the name so that it is highlighted blue and click on the tab from the bottom menu named “Available Downloads”.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 13.&#039;&#039;&#039; The available downloads tab and the 5 JSON files that are the QC outputs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
There will be 5 files reported in JSON format. Click the blue/green download icon next to each file to see the results. The file labeled &#039;&#039;&#039;qcAll.json&#039;&#039;&#039; will have our assemblyQC results. &#039;&#039;&#039;qcNGS.json&#039;&#039;&#039; will have our ngsQC results and &#039;&#039;&#039;biosample.json&#039;&#039;&#039; the biosample information. We currently do not submit qcPos.json or refAnnot.json to the ARGOS DB, but the information is there to better help you understand your computation.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_figure_2b.png&amp;diff=805</id>
		<title>File:Screenshot figure 2b.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_figure_2b.png&amp;diff=805"/>
		<updated>2025-04-25T19:57:05Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;fig&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_2a.png&amp;diff=804</id>
		<title>File:Screenshot fig 2a.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_fig_2a.png&amp;diff=804"/>
		<updated>2025-04-25T19:55:37Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;fig&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_figure_2_a.png&amp;diff=802</id>
		<title>File:Screenshot figure 2 a.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_figure_2_a.png&amp;diff=802"/>
		<updated>2025-04-25T19:53:07Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;fig&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=800</id>
		<title>ARGOSQC Usage Tutorial</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=800"/>
		<updated>2025-04-25T19:49:37Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
HIVE3 one-click pipeline tutorial for the FDA HIVE instance. This protocol will guide the user in running single and batch-mode QC computations. HIVE3 is an instance of HIVE not owned by the FDA which can be directly modified by Vahan or others on our team with permissions.&lt;br /&gt;
&lt;br /&gt;
= Required User Information =&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Protocol Version&#039;&#039;&#039; &lt;br /&gt;
|1.0&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Instance&#039;&#039;&#039;&lt;br /&gt;
|3&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Link&#039;&#039;&#039;&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=login&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Overview =&lt;br /&gt;
We constructed a QC one-click pipeline that takes user specified organism information and combines the 3 core ARGOS workflows to produce 5 different result datasets in JSON format (Figure 1). 3 out of the 5 result JSONs (assemblyQC, ngsQC, and biosampleMeta) have been updated and added to data.argosdb.&lt;br /&gt;
&lt;br /&gt;
To register your account, navigate to the link under the “Required User Information”. At the top right there will be a tab saying “register”. Fill out the appropriate fields and submit. Once submitted, please email mazumder_lab@gwu.edu so we can verify your account.&lt;br /&gt;
&lt;br /&gt;
The ARGOS pipeline can be accessed via the dropdown menu ‘Projects’ at the upper right hand screen and then under ‘Argos’. The pipeline in HIVE3 is located in the “Required User Information” section in the beginning of this protocol. &lt;br /&gt;
&lt;br /&gt;
After a successful login, you will be navigated to the home page. Use the menu at the top right corner under projects to access the ARGOS pipeline or use this URL to access the ARGOS QC pipeline on HIVE3:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=argos-alqc&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[File:Screenshot figure 1.png|left|thumb|723x723px|&#039;&#039;&#039;Figure 1.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline. ]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Input values:&#039;&#039; ==&lt;br /&gt;
On the ARGOS Pipeline input setting page, under the General tab, is where our data inputs for a single and batch computation will be taken. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Name:&#039;&#039;&#039; Give the computation a name. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Folder:&#039;&#039;&#039; Give the folder where your computations, data, and steps will be stored.&lt;br /&gt;
&lt;br /&gt;
* Can use: _ or -  or &amp;amp;  and letters + numbers&lt;br /&gt;
* Cannot use: / : ; , \ “ ” ‘ ’ &lt;br /&gt;
* Yet, you use / to create a sub folder, but that is not recommended. Manually moving the subfolder is best.&lt;br /&gt;
* Ex folder: Influenza A (h5n1)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reads:&#039;&#039;&#039; information needed for ngsQC &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;SRR&#039;&#039;&#039;: the SRR accession number, can be multiple per organism by using a “,” or populate extra fields by clicking on the gray + sign. This tool uses the NCBI SRA Fasterq function to grab the fastq files directly from NCBI without the user needing to import them to HIVE.  &lt;br /&gt;
* &#039;&#039;&#039;HIVE reads:&#039;&#039;&#039; Drop down menu can select reads already uploaded into HIVE, either from previous computations or manual uploads. &lt;br /&gt;
* See in Figure 2a&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the SRR input box, rather the HIVE IDs menu, but you can if you want to, it will just search within HIVE. See ngsQC Protocol for how to upload SRR information using external downloader. This external downloader process is the same as in HIVE2 and 1.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reference:&#039;&#039;&#039; Information used for the assemblyQC portion of the algorithm.&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Reference accession:&#039;&#039;&#039; This is the REFSEQ or Genbank accession number from NCBI or Genbank.&lt;br /&gt;
* &#039;&#039;&#039;Assembly ID&#039;&#039;&#039;: This is the ASSEMBLY accession number from NCBI. &lt;br /&gt;
* &#039;&#039;&#039;HIVE genome:&#039;&#039;&#039; Use the drop down menu to select a reference genome that has already been uploaded into HIVE. &lt;br /&gt;
* See in Figure 2b&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the assembly ID or reference accession input box, but you can if you want to. See AssemblyQC Protocol for how to upload assembly information using an external downloader or local upload. This process is the same as what is done in HIVE2.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Metadata: &#039;&#039;&#039; Used to grab information necessary to fill out the BiosampleMeta_HIVE document.&lt;br /&gt;
&lt;br /&gt;
* Biosample Accession: The optional accession number for the Biosample that was reported to be used when creating the assembly and will be linked to the SRR fastq files used for the ngsQC portion of the algorithm. This step is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Coding Table:&#039;&#039;&#039; Dropdown of genetic codon tables to be used for your computation, depending on the organism to be computed. The default is human, viral (Standard).&lt;br /&gt;
&lt;br /&gt;
* Tip: NCBI Taxonomy will list the codon table for each organism on their taxonomy page.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 2.&#039;&#039;&#039; The ARGOS_QC algorithm page input with all NCBI information for a test organism, Salmonella enterica. a) The &#039;&#039;&#039;SRR accession&#039;&#039;&#039; field contains both SRR fastqs for the organism that correspond to the biosample. &lt;br /&gt;
&lt;br /&gt;
b) The &#039;&#039;&#039;Reference Accession&#039;&#039;&#039; is the RefSeq Nucleotide accession number from NCBI, the &#039;&#039;&#039;Assembly ID&#039;&#039;&#039; is the assembly accession number from NCBI, and the Biosample Accession is the Biosample accession number from NCBI.&lt;br /&gt;
&lt;br /&gt;
a &amp;amp; b) &#039;&#039;&#039;HIVE IDs&#039;&#039;&#039; are accessions that are already uploaded into HIVE and the algorithm will automatically select this information as opposed to pulling information from outside sources. &lt;br /&gt;
&lt;br /&gt;
Almost all of this information can be found in the legacy assembly page for this organism, shown in &#039;&#039;&#039;Figure 3, 4, 5, 6, 7&#039;&#039;&#039;. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;a)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;b)&#039;&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Where to locate NCBI Information for the inputs:&#039;&#039; ==&lt;br /&gt;
Navigate to the NCBI legacy assembly page for your organism. Here you can find all of the information to be used for your computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 3.&#039;&#039;&#039; The information shown in the NCBI legacy assembly page. The RefSeq assembly accession corresponds to our organism of interest and will be used to fill out the &#039;&#039;&#039;“Assembly ID”&#039;&#039;&#039; input field on the HIVE3 ARGOS_QC input page. Please note that the bioproject matches the accession for the FDA_ARGOS bioproject, and there are 2 sequencing technologies listed, meaning that there will most likely be two SRR submissions that we can find on the SRA page (see Figure 5 and 6). &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 4&#039;&#039;&#039;. The bottom section of the legacy assembly page for our test organism. The column labeled RefSeq sequence lists the DNA RefSeq for our test organism which we will use for the “&#039;&#039;&#039;Reference Accession&#039;&#039;&#039;” field on the HIVE3 ARGOS_QC input page. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 5.&#039;&#039;&#039; Clicking on the Biosample accession number seen in Figure 3 will redirect you to the NCBI Biosample page for this organism. Under &#039;&#039;&#039;“Related Information”&#039;&#039;&#039;, click &#039;&#039;&#039;“SRA”&#039;&#039;&#039; to navigate to the NCBI SRA page for this biosample.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 6.&#039;&#039;&#039; The SRA page lists different sequencing links. Each link is reported to be sequenced on different platforms; either for Illumina or for the PacBio platform. This is common. The methodology behind using the different platforms is to gather insight for the assembly at different perspectives and levels. Illumina sequences DNA as multiple short reads that can be used to create an accurate reconstruction of the genomic sections analyzed by estimating the average/best fit nucleotide sequence. PacBio is a long read sequencer that takes “movies” of the DNA sequence as it moves through the technology and captures the sequence in one go from start to finish. The long read sequence then acts as a map for the short accurate short read sequences that need to be assembled. Therefore, it is important to use all of the links reported in our QC pipeline.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 7.&#039;&#039;&#039; Clicking on the bottom link under the &#039;&#039;&#039;Runs&#039;&#039;&#039; section from the SRA page shown in &#039;&#039;&#039;Figure 6&#039;&#039;&#039; will redirect to the page containing the SRR file. Copy and paste the SRR accession number into the input field in the HIVE3 pipeline labeled “SRR”. You will need to do this for both (or more) SRR accessions. Check that the bioproject, biosample, and organism name all coincide with our test organism.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Single QC Computation =&lt;br /&gt;
A single QC computation will allow for assemblyQC, biosampleQC, and ngsQC to be performed on one organism with one assembly, but can include multiple SRR ids. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Input the name of the computation and the name of the folder you wish to store the computation in. If you would like to add to a pre-existing folder, input the exact folder name in the folder input field. It is case sensitive. See input criteria at the beginning of this protocol.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; Under the dropdown menu for Reads, select which input you will use. &lt;br /&gt;
&lt;br /&gt;
* For SRRs, type or paste in the SRR id. If there is more than one SRR id, click on the gray + sign to populate a new input field or use a , .&lt;br /&gt;
** Troubleshooting: if the computation fails, try removing the spaces in between the commas and SRR ids. No spaces.&lt;br /&gt;
* For HIVE IDs, click on the HIVE ID option from the dropdown menu. Click on the gray dropdown menu arrow next to HIVE reads. A pop-up window will open, as seen in Figure 9. &lt;br /&gt;
** Click on the ids that you wish to use in the computation. Use Ctrl + shift to highlight multiple ids.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 8.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline filled out for a single QC computation of our test organism, Salmonella enterica. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 9.&#039;&#039;&#039; Pop-up window for SRR HIVE ID selection. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Next to Reference, select the input object you would like to use for the computation.&lt;br /&gt;
&lt;br /&gt;
* For Reference Accession, type or paste in the id you wish to use.&lt;br /&gt;
* For Assembly IDs, type or paste in the id you wish to use. &lt;br /&gt;
* For HIVE Genomes, refer to Step 2 above on how to select a HIVE id. It is the same process.&lt;br /&gt;
* Refer to the beginning of this protocol for what ids can be inputted.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Next under BioSample Accessions, paste in the biosample ID you would like to use for the computation. This is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Lastly, select from the Coding Table dropdown menu the genetic code you would like to use for your computation, if applicable. The default is “human, viral (Standard)”.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the information has been inputted correctly, click on the big blue Submit button in the middle, as seen in Figure 8.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 7:&#039;&#039;&#039; Once your computation has been submitted, you can click on the Home tab found in the top left to go back to the homepage and see the workflow.&lt;br /&gt;
&lt;br /&gt;
= Batch Mode Computation =&lt;br /&gt;
Batch mode operates by a user-specified ratio of groups. With the help of semicolons and commas, the ratio would be 1:1:1 for a batch mode computation. It is 1:1:1, because we are clustering them by ; so the pipeline recognizes the ids between the ; as one computation. It would be one cluster of SRRs to one assembly to one biosample, that is one computation. There is a colorful and highlighted example below displaying the syntax for the inputs.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;They would be grouped for computations like this example:&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
To separate between batches, use a semicolon “;” in between the IDs. A comma denotes separate IDs but semicolons as separate batches. These will be inputted in the General tab of the pipeline, same as single computation. Within each field, this is how the above example would look like in batch mode:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456, SRR0123457, SRR0123458, SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039; SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
Assembly IDs: GCA_0011223344.1&#039;&#039;&#039;;&#039;&#039;&#039; GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
BioSample Accessions: SAMN110654321, SAMN110654322&#039;&#039;&#039;;&#039;&#039;&#039; SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Notice the semicolon separation according to the example above. The commas separate the ids, and the semicolon the batch.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Troubleshooting Note:&#039;&#039;&#039; If your computation fails or if there is an error, remove the spaces between the , and ; .Previously, this has thrown an error but has been fixed, but worth a shot if your computation fails. It would look like:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456,SRR0123457,SRR0123458,SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039;SRR0123451,SRR0123452,SRR0123453,SRR0123454&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Navigate to the tab title Batch, Figure 10. This can be found on the ARGOS input settings page, Figure 1., next to the General tab. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; For the parameter “batch service&amp;quot; at the bottom select, from the dropdown menu batch mode. This will have the pipeline set to Batch Mode rather than single computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 10.&#039;&#039;&#039; Batch mode input settings window.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Selecting the parameters. Click on the drop down menu next to the text “Parameter list”.&lt;br /&gt;
&lt;br /&gt;
* Use the black plus button next to ‘Parameter List’ to populate an entry field.&lt;br /&gt;
* Select from the dropdown field the correct parameter based on the input field you used in the general input page. This can be seen in Figure 11.&lt;br /&gt;
* For example, if you pasted in SRR Ids you would choose the parameter SRR IDs. If you chose HIVEIDs you would select HIVE IDs from the dropdown.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 11.&#039;&#039;&#039; Dropdown menu from the Batch tab displaying the parameter options. Select the batch parameter option based on what the input information is in the general tab. For example, if you entered SRR ids, select SRR IDs from the dropdown. If you input a reference ID, select Reference IDs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Input the ratio for the batch service. &lt;br /&gt;
&lt;br /&gt;
* For computations in batch mode in the one-click pipeline, the computations are separated by semicolon “;” and the IDs within the computations by a comma “,”. Since the workflow will parse the computations and recognize the IDs between the “;” as one computation, the ratio will be 1:1:1. &lt;br /&gt;
* If the ratio is 1:1:1 then enter the value 1 for each box.&lt;br /&gt;
* One set of SRRs to one assembly to one biosample. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Inputting the information correctly in each field. Navigate back to the input settings page, Figure 1. The same page that you had used for the single computations will be used, but the only difference will be the semicolons and commas. The example below will visually show you how the information will be inputted for a batch mode.&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Batch Mode Parameter breakdown:&#039;&#039; ==&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Again, this is very similar to single computations, except that the batch mode will use semicolons and commas to separate the ids.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the input information is complete, hit the blue button Submit. You may exit the Argos pipeline window by hitting “Home” on the top left corner.&lt;br /&gt;
&lt;br /&gt;
= QC Computation Results =&lt;br /&gt;
Once you have submitted your computations, either single or batch, you should see the pipeline workflow in your Inbox or All Objects, as shown in Figure 12. You can also view the pipeline by clicking on the “workflows” tab also seen in Figure 12.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 12.&#039;&#039;&#039;  The pipeline workflow displayed in the user’s inbox.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
As the workflow progresses, your computations will be stored in the folder that you named from the beginning of this protocol. To view the contents of the folder, simply click on the plus sign next to the folder or the folder name to open.&lt;br /&gt;
&lt;br /&gt;
Once your computations are complete the QC outputs will be stored in JSON file format from the computation “&#039;&#039;&#039;Post-Alignment Quality Controls&#039;&#039;&#039;” or under the &#039;&#039;&#039;“CFlow”&#039;&#039;&#039; workflow. P-A QC can be found in the folder you specified for the computation or CFlow in All Objects. To view the JSONs click on the name so that it is highlighted blue and click on the tab from the bottom menu named “Available Downloads”.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 13.&#039;&#039;&#039; The available downloads tab and the 5 JSON files that are the QC outputs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
There will be 5 files reported in JSON format. Click the blue/green download icon next to each file to see the results. The file labeled &#039;&#039;&#039;qcAll.json&#039;&#039;&#039; will have our assemblyQC results. &#039;&#039;&#039;qcNGS.json&#039;&#039;&#039; will have our ngsQC results and &#039;&#039;&#039;biosample.json&#039;&#039;&#039; the biosample information. We currently do not submit qcPos.json or refAnnot.json to the ARGOS DB, but the information is there to better help you understand your computation.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=798</id>
		<title>ARGOSQC Usage Tutorial</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=798"/>
		<updated>2025-04-25T19:49:06Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: /* Overview */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
HIVE3 one-click pipeline tutorial for the FDA HIVE instance. This protocol will guide the user in running single and batch-mode QC computations. HIVE3 is an instance of HIVE not owned by the FDA which can be directly modified by Vahan or others on our team with permissions.&lt;br /&gt;
&lt;br /&gt;
= Required User Information =&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Protocol Version&#039;&#039;&#039; &lt;br /&gt;
|1.0&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Instance&#039;&#039;&#039;&lt;br /&gt;
|3&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Link&#039;&#039;&#039;&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=login&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Overview =&lt;br /&gt;
We constructed a QC one-click pipeline that takes user specified organism information and combines the 3 core ARGOS workflows to produce 5 different result datasets in JSON format (Figure 1). 3 out of the 5 result JSONs (assemblyQC, ngsQC, and biosampleMeta) have been updated and added to data.argosdb.&lt;br /&gt;
&lt;br /&gt;
To register your account, navigate to the link under the “Required User Information”. At the top right there will be a tab saying “register”. Fill out the appropriate fields and submit. Once submitted, please email mazumder_lab@gwu.edu so we can verify your account.&lt;br /&gt;
&lt;br /&gt;
The ARGOS pipeline can be accessed via the dropdown menu ‘Projects’ at the upper right hand screen and then under ‘Argos’. The pipeline in HIVE3 is located in the “Required User Information” section in the beginning of this protocol. &lt;br /&gt;
&lt;br /&gt;
After a successful login, you will be navigated to the home page. Use the menu at the top right corner under projects to access the ARGOS pipeline or use this URL to access the ARGOS QC pipeline on HIVE3:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=argos-alqc&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
[[File:Screenshot figure 1.png|left|thumb|723x723px|&#039;&#039;&#039;Figure 1.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline. ]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Input values:&#039;&#039; ==&lt;br /&gt;
On the ARGOS Pipeline input setting page, under the General tab, is where our data inputs for a single and batch computation will be taken. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Name:&#039;&#039;&#039; Give the computation a name. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Folder:&#039;&#039;&#039; Give the folder where your computations, data, and steps will be stored.&lt;br /&gt;
&lt;br /&gt;
* Can use: _ or -  or &amp;amp;  and letters + numbers&lt;br /&gt;
* Cannot use: / : ; , \ “ ” ‘ ’ &lt;br /&gt;
* Yet, you use / to create a sub folder, but that is not recommended. Manually moving the subfolder is best.&lt;br /&gt;
* Ex folder: Influenza A (h5n1)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reads:&#039;&#039;&#039; information needed for ngsQC &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;SRR&#039;&#039;&#039;: the SRR accession number, can be multiple per organism by using a “,” or populate extra fields by clicking on the gray + sign. This tool uses the NCBI SRA Fasterq function to grab the fastq files directly from NCBI without the user needing to import them to HIVE.  &lt;br /&gt;
* &#039;&#039;&#039;HIVE reads:&#039;&#039;&#039; Drop down menu can select reads already uploaded into HIVE, either from previous computations or manual uploads. &lt;br /&gt;
* See in Figure 2a&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the SRR input box, rather the HIVE IDs menu, but you can if you want to, it will just search within HIVE. See ngsQC Protocol for how to upload SRR information using external downloader. This external downloader process is the same as in HIVE2 and 1.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reference:&#039;&#039;&#039; Information used for the assemblyQC portion of the algorithm.&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Reference accession:&#039;&#039;&#039; This is the REFSEQ or Genbank accession number from NCBI or Genbank.&lt;br /&gt;
* &#039;&#039;&#039;Assembly ID&#039;&#039;&#039;: This is the ASSEMBLY accession number from NCBI. &lt;br /&gt;
* &#039;&#039;&#039;HIVE genome:&#039;&#039;&#039; Use the drop down menu to select a reference genome that has already been uploaded into HIVE. &lt;br /&gt;
* See in Figure 2b&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the assembly ID or reference accession input box, but you can if you want to. See AssemblyQC Protocol for how to upload assembly information using an external downloader or local upload. This process is the same as what is done in HIVE2.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Metadata: &#039;&#039;&#039; Used to grab information necessary to fill out the BiosampleMeta_HIVE document.&lt;br /&gt;
&lt;br /&gt;
* Biosample Accession: The optional accession number for the Biosample that was reported to be used when creating the assembly and will be linked to the SRR fastq files used for the ngsQC portion of the algorithm. This step is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Coding Table:&#039;&#039;&#039; Dropdown of genetic codon tables to be used for your computation, depending on the organism to be computed. The default is human, viral (Standard).&lt;br /&gt;
&lt;br /&gt;
* Tip: NCBI Taxonomy will list the codon table for each organism on their taxonomy page.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 2.&#039;&#039;&#039; The ARGOS_QC algorithm page input with all NCBI information for a test organism, Salmonella enterica. a) The &#039;&#039;&#039;SRR accession&#039;&#039;&#039; field contains both SRR fastqs for the organism that correspond to the biosample. &lt;br /&gt;
&lt;br /&gt;
b) The &#039;&#039;&#039;Reference Accession&#039;&#039;&#039; is the RefSeq Nucleotide accession number from NCBI, the &#039;&#039;&#039;Assembly ID&#039;&#039;&#039; is the assembly accession number from NCBI, and the Biosample Accession is the Biosample accession number from NCBI.&lt;br /&gt;
&lt;br /&gt;
a &amp;amp; b) &#039;&#039;&#039;HIVE IDs&#039;&#039;&#039; are accessions that are already uploaded into HIVE and the algorithm will automatically select this information as opposed to pulling information from outside sources. &lt;br /&gt;
&lt;br /&gt;
Almost all of this information can be found in the legacy assembly page for this organism, shown in &#039;&#039;&#039;Figure 3, 4, 5, 6, 7&#039;&#039;&#039;. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;a)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;b)&#039;&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Where to locate NCBI Information for the inputs:&#039;&#039; ==&lt;br /&gt;
Navigate to the NCBI legacy assembly page for your organism. Here you can find all of the information to be used for your computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 3.&#039;&#039;&#039; The information shown in the NCBI legacy assembly page. The RefSeq assembly accession corresponds to our organism of interest and will be used to fill out the &#039;&#039;&#039;“Assembly ID”&#039;&#039;&#039; input field on the HIVE3 ARGOS_QC input page. Please note that the bioproject matches the accession for the FDA_ARGOS bioproject, and there are 2 sequencing technologies listed, meaning that there will most likely be two SRR submissions that we can find on the SRA page (see Figure 5 and 6). &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 4&#039;&#039;&#039;. The bottom section of the legacy assembly page for our test organism. The column labeled RefSeq sequence lists the DNA RefSeq for our test organism which we will use for the “&#039;&#039;&#039;Reference Accession&#039;&#039;&#039;” field on the HIVE3 ARGOS_QC input page. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 5.&#039;&#039;&#039; Clicking on the Biosample accession number seen in Figure 3 will redirect you to the NCBI Biosample page for this organism. Under &#039;&#039;&#039;“Related Information”&#039;&#039;&#039;, click &#039;&#039;&#039;“SRA”&#039;&#039;&#039; to navigate to the NCBI SRA page for this biosample.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 6.&#039;&#039;&#039; The SRA page lists different sequencing links. Each link is reported to be sequenced on different platforms; either for Illumina or for the PacBio platform. This is common. The methodology behind using the different platforms is to gather insight for the assembly at different perspectives and levels. Illumina sequences DNA as multiple short reads that can be used to create an accurate reconstruction of the genomic sections analyzed by estimating the average/best fit nucleotide sequence. PacBio is a long read sequencer that takes “movies” of the DNA sequence as it moves through the technology and captures the sequence in one go from start to finish. The long read sequence then acts as a map for the short accurate short read sequences that need to be assembled. Therefore, it is important to use all of the links reported in our QC pipeline.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 7.&#039;&#039;&#039; Clicking on the bottom link under the &#039;&#039;&#039;Runs&#039;&#039;&#039; section from the SRA page shown in &#039;&#039;&#039;Figure 6&#039;&#039;&#039; will redirect to the page containing the SRR file. Copy and paste the SRR accession number into the input field in the HIVE3 pipeline labeled “SRR”. You will need to do this for both (or more) SRR accessions. Check that the bioproject, biosample, and organism name all coincide with our test organism.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Single QC Computation =&lt;br /&gt;
A single QC computation will allow for assemblyQC, biosampleQC, and ngsQC to be performed on one organism with one assembly, but can include multiple SRR ids. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Input the name of the computation and the name of the folder you wish to store the computation in. If you would like to add to a pre-existing folder, input the exact folder name in the folder input field. It is case sensitive. See input criteria at the beginning of this protocol.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; Under the dropdown menu for Reads, select which input you will use. &lt;br /&gt;
&lt;br /&gt;
* For SRRs, type or paste in the SRR id. If there is more than one SRR id, click on the gray + sign to populate a new input field or use a , .&lt;br /&gt;
** Troubleshooting: if the computation fails, try removing the spaces in between the commas and SRR ids. No spaces.&lt;br /&gt;
* For HIVE IDs, click on the HIVE ID option from the dropdown menu. Click on the gray dropdown menu arrow next to HIVE reads. A pop-up window will open, as seen in Figure 9. &lt;br /&gt;
** Click on the ids that you wish to use in the computation. Use Ctrl + shift to highlight multiple ids.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 8.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline filled out for a single QC computation of our test organism, Salmonella enterica. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 9.&#039;&#039;&#039; Pop-up window for SRR HIVE ID selection. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Next to Reference, select the input object you would like to use for the computation.&lt;br /&gt;
&lt;br /&gt;
* For Reference Accession, type or paste in the id you wish to use.&lt;br /&gt;
* For Assembly IDs, type or paste in the id you wish to use. &lt;br /&gt;
* For HIVE Genomes, refer to Step 2 above on how to select a HIVE id. It is the same process.&lt;br /&gt;
* Refer to the beginning of this protocol for what ids can be inputted.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Next under BioSample Accessions, paste in the biosample ID you would like to use for the computation. This is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Lastly, select from the Coding Table dropdown menu the genetic code you would like to use for your computation, if applicable. The default is “human, viral (Standard)”.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the information has been inputted correctly, click on the big blue Submit button in the middle, as seen in Figure 8.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 7:&#039;&#039;&#039; Once your computation has been submitted, you can click on the Home tab found in the top left to go back to the homepage and see the workflow.&lt;br /&gt;
&lt;br /&gt;
= Batch Mode Computation =&lt;br /&gt;
Batch mode operates by a user-specified ratio of groups. With the help of semicolons and commas, the ratio would be 1:1:1 for a batch mode computation. It is 1:1:1, because we are clustering them by ; so the pipeline recognizes the ids between the ; as one computation. It would be one cluster of SRRs to one assembly to one biosample, that is one computation. There is a colorful and highlighted example below displaying the syntax for the inputs.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;They would be grouped for computations like this example:&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
To separate between batches, use a semicolon “;” in between the IDs. A comma denotes separate IDs but semicolons as separate batches. These will be inputted in the General tab of the pipeline, same as single computation. Within each field, this is how the above example would look like in batch mode:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456, SRR0123457, SRR0123458, SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039; SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
Assembly IDs: GCA_0011223344.1&#039;&#039;&#039;;&#039;&#039;&#039; GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
BioSample Accessions: SAMN110654321, SAMN110654322&#039;&#039;&#039;;&#039;&#039;&#039; SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Notice the semicolon separation according to the example above. The commas separate the ids, and the semicolon the batch.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Troubleshooting Note:&#039;&#039;&#039; If your computation fails or if there is an error, remove the spaces between the , and ; .Previously, this has thrown an error but has been fixed, but worth a shot if your computation fails. It would look like:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456,SRR0123457,SRR0123458,SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039;SRR0123451,SRR0123452,SRR0123453,SRR0123454&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Navigate to the tab title Batch, Figure 10. This can be found on the ARGOS input settings page, Figure 1., next to the General tab. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; For the parameter “batch service&amp;quot; at the bottom select, from the dropdown menu batch mode. This will have the pipeline set to Batch Mode rather than single computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 10.&#039;&#039;&#039; Batch mode input settings window.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Selecting the parameters. Click on the drop down menu next to the text “Parameter list”.&lt;br /&gt;
&lt;br /&gt;
* Use the black plus button next to ‘Parameter List’ to populate an entry field.&lt;br /&gt;
* Select from the dropdown field the correct parameter based on the input field you used in the general input page. This can be seen in Figure 11.&lt;br /&gt;
* For example, if you pasted in SRR Ids you would choose the parameter SRR IDs. If you chose HIVEIDs you would select HIVE IDs from the dropdown.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 11.&#039;&#039;&#039; Dropdown menu from the Batch tab displaying the parameter options. Select the batch parameter option based on what the input information is in the general tab. For example, if you entered SRR ids, select SRR IDs from the dropdown. If you input a reference ID, select Reference IDs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Input the ratio for the batch service. &lt;br /&gt;
&lt;br /&gt;
* For computations in batch mode in the one-click pipeline, the computations are separated by semicolon “;” and the IDs within the computations by a comma “,”. Since the workflow will parse the computations and recognize the IDs between the “;” as one computation, the ratio will be 1:1:1. &lt;br /&gt;
* If the ratio is 1:1:1 then enter the value 1 for each box.&lt;br /&gt;
* One set of SRRs to one assembly to one biosample. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Inputting the information correctly in each field. Navigate back to the input settings page, Figure 1. The same page that you had used for the single computations will be used, but the only difference will be the semicolons and commas. The example below will visually show you how the information will be inputted for a batch mode.&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Batch Mode Parameter breakdown:&#039;&#039; ==&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Again, this is very similar to single computations, except that the batch mode will use semicolons and commas to separate the ids.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the input information is complete, hit the blue button Submit. You may exit the Argos pipeline window by hitting “Home” on the top left corner.&lt;br /&gt;
&lt;br /&gt;
= QC Computation Results =&lt;br /&gt;
Once you have submitted your computations, either single or batch, you should see the pipeline workflow in your Inbox or All Objects, as shown in Figure 12. You can also view the pipeline by clicking on the “workflows” tab also seen in Figure 12.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 12.&#039;&#039;&#039;  The pipeline workflow displayed in the user’s inbox.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
As the workflow progresses, your computations will be stored in the folder that you named from the beginning of this protocol. To view the contents of the folder, simply click on the plus sign next to the folder or the folder name to open.&lt;br /&gt;
&lt;br /&gt;
Once your computations are complete the QC outputs will be stored in JSON file format from the computation “&#039;&#039;&#039;Post-Alignment Quality Controls&#039;&#039;&#039;” or under the &#039;&#039;&#039;“CFlow”&#039;&#039;&#039; workflow. P-A QC can be found in the folder you specified for the computation or CFlow in All Objects. To view the JSONs click on the name so that it is highlighted blue and click on the tab from the bottom menu named “Available Downloads”.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 13.&#039;&#039;&#039; The available downloads tab and the 5 JSON files that are the QC outputs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
There will be 5 files reported in JSON format. Click the blue/green download icon next to each file to see the results. The file labeled &#039;&#039;&#039;qcAll.json&#039;&#039;&#039; will have our assemblyQC results. &#039;&#039;&#039;qcNGS.json&#039;&#039;&#039; will have our ngsQC results and &#039;&#039;&#039;biosample.json&#039;&#039;&#039; the biosample information. We currently do not submit qcPos.json or refAnnot.json to the ARGOS DB, but the information is there to better help you understand your computation.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_figure_1.png&amp;diff=797</id>
		<title>File:Screenshot figure 1.png</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=File:Screenshot_figure_1.png&amp;diff=797"/>
		<updated>2025-04-25T19:45:26Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Figure 1 for th epipeline&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
	<entry>
		<id>https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=796</id>
		<title>ARGOSQC Usage Tutorial</title>
		<link rel="alternate" type="text/html" href="https://hivelab.biochemistry.gwu.edu/wiki/index.php?title=ARGOSQC_Usage_Tutorial&amp;diff=796"/>
		<updated>2025-04-25T19:43:13Z</updated>

		<summary type="html">&lt;p&gt;Christie.woodside: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
HIVE3 one-click pipeline tutorial for the FDA HIVE instance. This protocol will guide the user in running single and batch-mode QC computations. HIVE3 is an instance of HIVE not owned by the FDA which can be directly modified by Vahan or others on our team with permissions.&lt;br /&gt;
&lt;br /&gt;
= Required User Information =&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Protocol Version&#039;&#039;&#039; &lt;br /&gt;
|1.0&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Instance&#039;&#039;&#039;&lt;br /&gt;
|3&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;HIVE Link&#039;&#039;&#039;&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=login&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Overview =&lt;br /&gt;
We constructed a QC one-click pipeline that takes user specified organism information and combines the 3 core ARGOS workflows to produce 5 different result datasets in JSON format (Figure 1). 3 out of the 5 result JSONs (assemblyQC, ngsQC, and biosampleMeta) have been updated and added to data.argosdb.&lt;br /&gt;
&lt;br /&gt;
To register your account, navigate to the link under the “Required User Information”. At the top right there will be a tab saying “register”. Fill out the appropriate fields and submit. Once submitted, please email mazumder_lab@gwu.edu so we can verify your account.&lt;br /&gt;
&lt;br /&gt;
The ARGOS pipeline can be accessed via the dropdown menu ‘Projects’ at the upper right hand screen and then under ‘Argos’. The pipeline in HIVE3 is located in the “Required User Information” section in the beginning of this protocol. &lt;br /&gt;
&lt;br /&gt;
After a successful login, you will be navigated to the home page. Use the menu at the top right corner under projects to access the ARGOS pipeline or use this URL to access the ARGOS QC pipeline on HIVE3:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;nowiki&amp;gt;https://hive3.biochemistry.gwu.edu/dna.cgi?cmd=argos-alqc&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 1.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Input values:&#039;&#039; ==&lt;br /&gt;
On the ARGOS Pipeline input setting page, under the General tab, is where our data inputs for a single and batch computation will be taken. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Name:&#039;&#039;&#039; Give the computation a name. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Folder:&#039;&#039;&#039; Give the folder where your computations, data, and steps will be stored.&lt;br /&gt;
&lt;br /&gt;
* Can use: _ or -  or &amp;amp;  and letters + numbers&lt;br /&gt;
* Cannot use: / : ; , \ “ ” ‘ ’ &lt;br /&gt;
* Yet, you use / to create a sub folder, but that is not recommended. Manually moving the subfolder is best.&lt;br /&gt;
* Ex folder: Influenza A (h5n1)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reads:&#039;&#039;&#039; information needed for ngsQC &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;SRR&#039;&#039;&#039;: the SRR accession number, can be multiple per organism by using a “,” or populate extra fields by clicking on the gray + sign. This tool uses the NCBI SRA Fasterq function to grab the fastq files directly from NCBI without the user needing to import them to HIVE.  &lt;br /&gt;
* &#039;&#039;&#039;HIVE reads:&#039;&#039;&#039; Drop down menu can select reads already uploaded into HIVE, either from previous computations or manual uploads. &lt;br /&gt;
* See in Figure 2a&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the SRR input box, rather the HIVE IDs menu, but you can if you want to, it will just search within HIVE. See ngsQC Protocol for how to upload SRR information using external downloader. This external downloader process is the same as in HIVE2 and 1.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Reference:&#039;&#039;&#039; Information used for the assemblyQC portion of the algorithm.&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Reference accession:&#039;&#039;&#039; This is the REFSEQ or Genbank accession number from NCBI or Genbank.&lt;br /&gt;
* &#039;&#039;&#039;Assembly ID&#039;&#039;&#039;: This is the ASSEMBLY accession number from NCBI. &lt;br /&gt;
* &#039;&#039;&#039;HIVE genome:&#039;&#039;&#039; Use the drop down menu to select a reference genome that has already been uploaded into HIVE. &lt;br /&gt;
* See in Figure 2b&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;&#039;&#039;NOTE:&#039;&#039;&#039;&#039;&#039; The algorithm will automatically select information already in HIVE as opposed to pulling information from outside sources. If the reads are already uploaded you do not need to use the assembly ID or reference accession input box, but you can if you want to. See AssemblyQC Protocol for how to upload assembly information using an external downloader or local upload. This process is the same as what is done in HIVE2.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Metadata: &#039;&#039;&#039; Used to grab information necessary to fill out the BiosampleMeta_HIVE document.&lt;br /&gt;
&lt;br /&gt;
* Biosample Accession: The optional accession number for the Biosample that was reported to be used when creating the assembly and will be linked to the SRR fastq files used for the ngsQC portion of the algorithm. This step is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Coding Table:&#039;&#039;&#039; Dropdown of genetic codon tables to be used for your computation, depending on the organism to be computed. The default is human, viral (Standard).&lt;br /&gt;
&lt;br /&gt;
* Tip: NCBI Taxonomy will list the codon table for each organism on their taxonomy page.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 2.&#039;&#039;&#039; The ARGOS_QC algorithm page input with all NCBI information for a test organism, Salmonella enterica. a) The &#039;&#039;&#039;SRR accession&#039;&#039;&#039; field contains both SRR fastqs for the organism that correspond to the biosample. &lt;br /&gt;
&lt;br /&gt;
b) The &#039;&#039;&#039;Reference Accession&#039;&#039;&#039; is the RefSeq Nucleotide accession number from NCBI, the &#039;&#039;&#039;Assembly ID&#039;&#039;&#039; is the assembly accession number from NCBI, and the Biosample Accession is the Biosample accession number from NCBI.&lt;br /&gt;
&lt;br /&gt;
a &amp;amp; b) &#039;&#039;&#039;HIVE IDs&#039;&#039;&#039; are accessions that are already uploaded into HIVE and the algorithm will automatically select this information as opposed to pulling information from outside sources. &lt;br /&gt;
&lt;br /&gt;
Almost all of this information can be found in the legacy assembly page for this organism, shown in &#039;&#039;&#039;Figure 3, 4, 5, 6, 7&#039;&#039;&#039;. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;a)&#039;&#039;&#039;&lt;br /&gt;
|-&lt;br /&gt;
|&#039;&#039;&#039;b)&#039;&#039;&#039;&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Where to locate NCBI Information for the inputs:&#039;&#039; ==&lt;br /&gt;
Navigate to the NCBI legacy assembly page for your organism. Here you can find all of the information to be used for your computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 3.&#039;&#039;&#039; The information shown in the NCBI legacy assembly page. The RefSeq assembly accession corresponds to our organism of interest and will be used to fill out the &#039;&#039;&#039;“Assembly ID”&#039;&#039;&#039; input field on the HIVE3 ARGOS_QC input page. Please note that the bioproject matches the accession for the FDA_ARGOS bioproject, and there are 2 sequencing technologies listed, meaning that there will most likely be two SRR submissions that we can find on the SRA page (see Figure 5 and 6). &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 4&#039;&#039;&#039;. The bottom section of the legacy assembly page for our test organism. The column labeled RefSeq sequence lists the DNA RefSeq for our test organism which we will use for the “&#039;&#039;&#039;Reference Accession&#039;&#039;&#039;” field on the HIVE3 ARGOS_QC input page. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 5.&#039;&#039;&#039; Clicking on the Biosample accession number seen in Figure 3 will redirect you to the NCBI Biosample page for this organism. Under &#039;&#039;&#039;“Related Information”&#039;&#039;&#039;, click &#039;&#039;&#039;“SRA”&#039;&#039;&#039; to navigate to the NCBI SRA page for this biosample.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 6.&#039;&#039;&#039; The SRA page lists different sequencing links. Each link is reported to be sequenced on different platforms; either for Illumina or for the PacBio platform. This is common. The methodology behind using the different platforms is to gather insight for the assembly at different perspectives and levels. Illumina sequences DNA as multiple short reads that can be used to create an accurate reconstruction of the genomic sections analyzed by estimating the average/best fit nucleotide sequence. PacBio is a long read sequencer that takes “movies” of the DNA sequence as it moves through the technology and captures the sequence in one go from start to finish. The long read sequence then acts as a map for the short accurate short read sequences that need to be assembled. Therefore, it is important to use all of the links reported in our QC pipeline.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 7.&#039;&#039;&#039; Clicking on the bottom link under the &#039;&#039;&#039;Runs&#039;&#039;&#039; section from the SRA page shown in &#039;&#039;&#039;Figure 6&#039;&#039;&#039; will redirect to the page containing the SRR file. Copy and paste the SRR accession number into the input field in the HIVE3 pipeline labeled “SRR”. You will need to do this for both (or more) SRR accessions. Check that the bioproject, biosample, and organism name all coincide with our test organism.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
= Single QC Computation =&lt;br /&gt;
A single QC computation will allow for assemblyQC, biosampleQC, and ngsQC to be performed on one organism with one assembly, but can include multiple SRR ids. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Input the name of the computation and the name of the folder you wish to store the computation in. If you would like to add to a pre-existing folder, input the exact folder name in the folder input field. It is case sensitive. See input criteria at the beginning of this protocol.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; Under the dropdown menu for Reads, select which input you will use. &lt;br /&gt;
&lt;br /&gt;
* For SRRs, type or paste in the SRR id. If there is more than one SRR id, click on the gray + sign to populate a new input field or use a , .&lt;br /&gt;
** Troubleshooting: if the computation fails, try removing the spaces in between the commas and SRR ids. No spaces.&lt;br /&gt;
* For HIVE IDs, click on the HIVE ID option from the dropdown menu. Click on the gray dropdown menu arrow next to HIVE reads. A pop-up window will open, as seen in Figure 9. &lt;br /&gt;
** Click on the ids that you wish to use in the computation. Use Ctrl + shift to highlight multiple ids.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 8.&#039;&#039;&#039; Input settings page for the ARGOS QC pipeline filled out for a single QC computation of our test organism, Salmonella enterica. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 9.&#039;&#039;&#039; Pop-up window for SRR HIVE ID selection. &lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Next to Reference, select the input object you would like to use for the computation.&lt;br /&gt;
&lt;br /&gt;
* For Reference Accession, type or paste in the id you wish to use.&lt;br /&gt;
* For Assembly IDs, type or paste in the id you wish to use. &lt;br /&gt;
* For HIVE Genomes, refer to Step 2 above on how to select a HIVE id. It is the same process.&lt;br /&gt;
* Refer to the beginning of this protocol for what ids can be inputted.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Next under BioSample Accessions, paste in the biosample ID you would like to use for the computation. This is optional.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Lastly, select from the Coding Table dropdown menu the genetic code you would like to use for your computation, if applicable. The default is “human, viral (Standard)”.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the information has been inputted correctly, click on the big blue Submit button in the middle, as seen in Figure 8.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 7:&#039;&#039;&#039; Once your computation has been submitted, you can click on the Home tab found in the top left to go back to the homepage and see the workflow.&lt;br /&gt;
&lt;br /&gt;
= Batch Mode Computation =&lt;br /&gt;
Batch mode operates by a user-specified ratio of groups. With the help of semicolons and commas, the ratio would be 1:1:1 for a batch mode computation. It is 1:1:1, because we are clustering them by ; so the pipeline recognizes the ids between the ; as one computation. It would be one cluster of SRRs to one assembly to one biosample, that is one computation. There is a colorful and highlighted example below displaying the syntax for the inputs.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;They would be grouped for computations like this example:&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
To separate between batches, use a semicolon “;” in between the IDs. A comma denotes separate IDs but semicolons as separate batches. These will be inputted in the General tab of the pipeline, same as single computation. Within each field, this is how the above example would look like in batch mode:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456, SRR0123457, SRR0123458, SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039; SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
Assembly IDs: GCA_0011223344.1&#039;&#039;&#039;;&#039;&#039;&#039; GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
BioSample Accessions: SAMN110654321, SAMN110654322&#039;&#039;&#039;;&#039;&#039;&#039; SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Notice the semicolon separation according to the example above. The commas separate the ids, and the semicolon the batch.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Troubleshooting Note:&#039;&#039;&#039; If your computation fails or if there is an error, remove the spaces between the , and ; .Previously, this has thrown an error but has been fixed, but worth a shot if your computation fails. It would look like:&lt;br /&gt;
&lt;br /&gt;
SRR IDS: SRR0123456,SRR0123457,SRR0123458,SRR0123459&#039;&#039;&#039;;&#039;&#039;&#039;SRR0123451,SRR0123452,SRR0123453,SRR0123454&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 1:&#039;&#039;&#039; Navigate to the tab title Batch, Figure 10. This can be found on the ARGOS input settings page, Figure 1., next to the General tab. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 2:&#039;&#039;&#039; For the parameter “batch service&amp;quot; at the bottom select, from the dropdown menu batch mode. This will have the pipeline set to Batch Mode rather than single computations.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 10.&#039;&#039;&#039; Batch mode input settings window.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 3:&#039;&#039;&#039; Selecting the parameters. Click on the drop down menu next to the text “Parameter list”.&lt;br /&gt;
&lt;br /&gt;
* Use the black plus button next to ‘Parameter List’ to populate an entry field.&lt;br /&gt;
* Select from the dropdown field the correct parameter based on the input field you used in the general input page. This can be seen in Figure 11.&lt;br /&gt;
* For example, if you pasted in SRR Ids you would choose the parameter SRR IDs. If you chose HIVEIDs you would select HIVE IDs from the dropdown.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 11.&#039;&#039;&#039; Dropdown menu from the Batch tab displaying the parameter options. Select the batch parameter option based on what the input information is in the general tab. For example, if you entered SRR ids, select SRR IDs from the dropdown. If you input a reference ID, select Reference IDs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&#039;&#039;&#039;Step 4:&#039;&#039;&#039; Input the ratio for the batch service. &lt;br /&gt;
&lt;br /&gt;
* For computations in batch mode in the one-click pipeline, the computations are separated by semicolon “;” and the IDs within the computations by a comma “,”. Since the workflow will parse the computations and recognize the IDs between the “;” as one computation, the ratio will be 1:1:1. &lt;br /&gt;
* If the ratio is 1:1:1 then enter the value 1 for each box.&lt;br /&gt;
* One set of SRRs to one assembly to one biosample. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 5:&#039;&#039;&#039; Inputting the information correctly in each field. Navigate back to the input settings page, Figure 1. The same page that you had used for the single computations will be used, but the only difference will be the semicolons and commas. The example below will visually show you how the information will be inputted for a batch mode.&lt;br /&gt;
&lt;br /&gt;
== &#039;&#039;Batch Mode Parameter breakdown:&#039;&#039; ==&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Assembly 1: &lt;br /&gt;
&lt;br /&gt;
SRR0123456, SRR0123457, SRR0123458, SRR0123459&lt;br /&gt;
&lt;br /&gt;
GCA_0011223344.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654321, SAMN110654322&lt;br /&gt;
&lt;br /&gt;
Assembly 2: &lt;br /&gt;
&lt;br /&gt;
SRR0123451, SRR0123452, SRR0123453, SRR0123454&lt;br /&gt;
&lt;br /&gt;
GCA_0011223345.1&lt;br /&gt;
&lt;br /&gt;
SAMN110654323, SAMN110654324&lt;br /&gt;
&lt;br /&gt;
Again, this is very similar to single computations, except that the batch mode will use semicolons and commas to separate the ids.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Step 6:&#039;&#039;&#039; Once all of the input information is complete, hit the blue button Submit. You may exit the Argos pipeline window by hitting “Home” on the top left corner.&lt;br /&gt;
&lt;br /&gt;
= QC Computation Results =&lt;br /&gt;
Once you have submitted your computations, either single or batch, you should see the pipeline workflow in your Inbox or All Objects, as shown in Figure 12. You can also view the pipeline by clicking on the “workflows” tab also seen in Figure 12.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 12.&#039;&#039;&#039;  The pipeline workflow displayed in the user’s inbox.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
As the workflow progresses, your computations will be stored in the folder that you named from the beginning of this protocol. To view the contents of the folder, simply click on the plus sign next to the folder or the folder name to open.&lt;br /&gt;
&lt;br /&gt;
Once your computations are complete the QC outputs will be stored in JSON file format from the computation “&#039;&#039;&#039;Post-Alignment Quality Controls&#039;&#039;&#039;” or under the &#039;&#039;&#039;“CFlow”&#039;&#039;&#039; workflow. P-A QC can be found in the folder you specified for the computation or CFlow in All Objects. To view the JSONs click on the name so that it is highlighted blue and click on the tab from the bottom menu named “Available Downloads”.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|&#039;&#039;&#039;Figure 13.&#039;&#039;&#039; The available downloads tab and the 5 JSON files that are the QC outputs.&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
There will be 5 files reported in JSON format. Click the blue/green download icon next to each file to see the results. The file labeled &#039;&#039;&#039;qcAll.json&#039;&#039;&#039; will have our assemblyQC results. &#039;&#039;&#039;qcNGS.json&#039;&#039;&#039; will have our ngsQC results and &#039;&#039;&#039;biosample.json&#039;&#039;&#039; the biosample information. We currently do not submit qcPos.json or refAnnot.json to the ARGOS DB, but the information is there to better help you understand your computation.&lt;/div&gt;</summary>
		<author><name>Christie.woodside</name></author>
	</entry>
</feed>