Swoogle Home



Login Form

Lost Password?
No account yet? Register

BSQ Sitestats Summary

Site Stats Summary
  Hits Visitors
Today 2914 396
Week 22760 2091

Swoogle Manual


Shortcuts:    General    How-To    Implementation    Research


Q: What does Swoogle do?

Swoogle is a search engine for the Semantic Web on the Web. Swoogle crawl the World Wide Web for a special class of web documents called Semantic Web documents, which are written in RDF. Currently, it provides the following services to the following services:

  • search Semantic Web ontologies
  • search Semantic Web instance data
  • search Semantic Web terms, i.e., URIs that have been defined as classes and properties
  • provide metadata of Semantic Web documents and support browsing the Semantic Web. (Please refer to Li Ding et. al., Finding and Ranking Knowledge on the Semantic Web, ISWC'04 for details)
  • archive different versions of Semantic Web documents

Currently, Swoogle only indexes some metadata about Semantic Web documents. It neither stores nor searches all triples in an Semantic Web documents as a triple store.

Q: Why do you name your system Swoogle?

Swoogle stands for "Semantic Web Ontology ..." Well, we're still trying to figure out the rest.

Q: Who is behind Swoogle?

Swoogle is a research project being carried out by the ebiquity research group in the Computer Science and Electrical Engineering Department at the University of Maryland, Baltimore County (UMBC). Partial research support was provided by DARPA contract F30602-00-0591 and by NSF by awards NSF-ITR-IIS-0326460 and NSF-ITR-IDM-0219649. Contributors include Tim Finin, Li Ding, Rong Pan, Anupam Joshi, Pavan Reddivari, Joel Sachs, Pranam Kolari, Akshay Java, Lushan Han, Yun Peng, R. Scott Cost,   Sandor Dornbush and Vishal Doshi.

Q: Does Swoogle provide web services?

Yes, Swoogle provide REST web service to the public users. Please read Search Service section in Swoogle Manual to learn the web service interface. In fact, the entire Swoogle website is based on the web services as well.

Q: Why should I registered and become a user?

Swoogle provide human friendly interface to public users. An anonymous user can only access at most 100 results for each query; however,  a registered user can browse much more results. If you are using the machine interface, you may contact Swoogle developers for a special account.

Q: Do you have any plans to commercialize Swoogle?

No. Swoogle is a research project. We have no interest in commercializing the ideas or technology.

Q: Can I get the source code?

Not at the moment. When Swoogle is more complete and stable, we may make it available under a suitable open source license.

Q: What kind of data does Swoogle index?

Swoogle index only Semantic Web documents, currently including those written in RDF/XML, N-Triples, N3(RDF)  and some documents that embed RDF/XML fragments. The data presented at Swoogle are collected from the public accessible World Wide Web. Swoogle has a privacy policy in English and a crawling policy.


Q: How to get help when I have problems or questions on using Swoogle? 

Please try Swoogle Manual first. You can join the Google group for Swoogle users, Swooglers@googlegroups.com, and then post you question to the group. Maybe other users will be able to help. If not, you can also try asking the Swoogle Development team by sending mail to swoogle-developers@cs.umbc.edu.

Q: How to send my feedback on Swoogle?

We're always happy to hear comments, suggestions and feedback. Send any thoughts you have to swoogle-developers@cs.umbc.edu.

Q: How to get back to the old Swoogle?

This version replaces the old version; however, we still maintain the old Swoogle (v2.2) and call it Swoogle 2005. It has been linked from the front page, or you may follow this link http://swoogle.umbc.edu/2005.

Q: How to submit URL and how to tell if my URLs have been indexed by Swoogle?

Swoogle provides the Semantic Web archive service to help users (i) check if a URL has been indexed (ii) track the previous versions of the Semantic Web document retrieved from the URL.  If you want to submit a new URL or a web page containing hyperlinks to your Semantic Web documents, you may directly go to submit URL page. Swoogle will run regular crawling starting from the provided URL.

Q: How to remove my URLs from Swoogle

We update our index on a regular basis, so updated or outdated URL submissions are not necessary. Documents that are not accessible will eventually be removed from the database. If you ever need to ask us to remove entries from the database, contact swoogle-developers@cs.umbc.edu.

Q: How to cite Swoogle?

The classic Swoogle Paper:

Li Ding, Tim Finin, Anupam Joshi, Rong Pan, R. Scott Cost, Yun Peng, Pavan Reddivari, Vishal C Doshi, and Joel Sachs, "Swoogle: A Search and Metadata Engine for the Semantic Web", In the Proceedings of the Thirteenth ACM Conference on Information and Knowledge Management , November 2004  [PDF] [BibTex].

The latest Swoogle Paper:

Li Ding, Rong Pan, Tim Finin, Anupam Joshi, Yun Peng, and Pranam Kolari, "Finding and Ranking Knowledge on the Semantic Web", in the Proceedings of the 4th International Semantic Web Conference, November 2005 [PDF] [BibTex]


Q: How does Swoogle crawl the semantic web?

Swoogle adopts a hybrid approach to harvest the Semantic Web, including manual submission, Google-based meta-crawling, bounded HTML crawling and RDF crawling.

Q: How many documents are indexed by Swoogle?

Swoogle's statistic page is under re-construction now. We have indexed over 1.4 million Semantic Web documents and 290 million triples, and daily updates can be found at this page.

Q: How Swoogle is implemented?

Swoogle is mainly written in Java (JDK 1.4.2), and its web services are provided through Apache Tomcat server. The front-end Swoogle website is written in PHP. Currently, Swoogle stores its data in a MySQL database (mysql 4.1.16). The entire system is running on a Linux platform (Fedor Core 4).

Q: Does Swoogle use any Triple store?

Swoogle is designed as a repository of URLs but not a triple store, so it does not store all triples encountered. However, we may build applications on top of Swoogle that index a specific class of Semantic Web data and provide inference support using triple stores.


Q: Where can I find information on the research relating to Swoogle?

You can find formal research papers on Swoogle at the project web-site .

Q: Are their any other systems like Swoogle?

Here's a partial list of related projects as of May 2006. Please let us know if there are others that aren't on the list.

  • manually maintained ontology repositories
  • Semantic Web search and browsers
    • Ontaria  offers a searchable and browsable directory of semantic web data based on triple store. It is developed by the W3C.
    • Semantic Web Search lets users search Semantic Web instances data. It is developed by IntelliDimension
    • ontoSearch, a Google based ontology search engine. It is developed by AKT.
  • crawlers
    • DAML Crawler is a program that collects DAML statements by traversing WWW references and links.
    • RDF Crawler is a Java crawler for RDF written in 2001, but the specification is recently updated in 2005
    • OCRA (Ontology CRAwler) is a simple Java crawler using Jena that uses seeAlso references and URI names.
    • scutter, another name of RDF crawler, from FOAF project

manual  o   news  o   faq  o   web-service  o   submit-url  o   sw-archive  o   feedback  o   swoogle2005

Swoogle © 2004-2007, ebiquity group at UMBC
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 2.5 License.