What you get
Cosmos (and its ecosystem) is the BigData Analysis Generic Enabler reference implementation (GEri).
The BigData Analysis GE is intended to deploy means for analyzing both batch and stream data (in order to get, in the end, insights on such a data). Although the streaming part is still in the roadmp, the batch part has been widely developed throught the adoption and/or the in-house creation of the following tools:
- A Hadoop As A Service (HAAS) engine, either the ''official'' one based on Openstack's Sahara, either the light version based on a shared Hadoop cluster.
- Cosmos GUI and an OAuth2 Tokens Generator for Cosmos REST APIs.
- Cygnus, the data connector for Orion Context Broker.
- Tidoop, some extensions for Hadoop including means of using CKAN data and some general purpose MapReduce jobs
Nevertheless, and according to the current versions of the software, not all the tools are mandatory; please have a look on the description of each tool in order to know wether it suits for your deployment or not.
Why to get it
Cosmos is mainly addressed to those service providers aiming to expose a BigData Analysis GE-like services. For those service providers, the data analysis is not a goal itself but providing ways others can perform such data analysis. This especially applies to Openstack's Sahara installation.
If you are a data scientist willing to get some insights on certain data; or you are a software engineer in charge of productizing an application based on a previous data scientist analysis, then please visit the User and Programmer Guide; and/or go directly to the FIWARE Lab global instance of Cosmos, there you will find an already deployed infrastructure ready to be used through the different APIs.
If you don't relay on FIWARE Lab global instance of Cosmos and you want to use Hadoop, do not install Cosmos; that will be as installing a complete Cloud just for creating a single virtual machine. Instead, simply install a private instance of Hadoop!
If you still have doubts, we have built this flow diagram below in order to help you identifying which kind of Big Data user you are (if any).