enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Apache Hadoop - Wikipedia

    en.wikipedia.org/wiki/Apache_Hadoop

    The term Hadoop is often used for both base modules and sub-modules and also the ecosystem, [12] or collection of additional software packages that can be installed on top of or alongside Hadoop, such as Apache Pig, Apache Hive, Apache HBase, Apache Phoenix, Apache Spark, Apache ZooKeeper, Apache Impala, Apache Flume, Apache Sqoop, Apache Oozie ...

  3. Apache Hive - Wikipedia

    en.wikipedia.org/wiki/Apache_Hive

    Apache Hive is a data warehouse software project. It is built on top of Apache Hadoop for providing data query and analysis. [3] [4] Hive gives an SQL-like interface to query data stored in various databases and file systems that integrate with Hadoop.

  4. List of Apache Software Foundation projects - Wikipedia

    en.wikipedia.org/wiki/List_of_Apache_Software...

    Kibble: a suite of tools for collecting, aggregating and visualizing activity in software projects. Knox: a REST API Gateway for Hadoop Services; Kudu: a distributed columnar storage engine built for the Apache Hadoop ecosystem; Kvrocks: a distributed key-value NoSQL database, supporting the rich data structure; Kylin: distributed analytics engine

  5. Apache HBase - Wikipedia

    en.wikipedia.org/wiki/Apache_HBase

    Tables in HBase can serve as the input and output for MapReduce jobs run in Hadoop, and may be accessed through the Java API but also through REST, Avro or Thrift gateway APIs. HBase is a wide-column store and has been widely adopted because of its lineage with Hadoop and HDFS. HBase runs on top of HDFS and is well-suited for fast read and ...

  6. Apache Avro - Wikipedia

    en.wikipedia.org/wiki/Apache_Avro

    Avro is a row-oriented remote procedure call and data serialization framework developed within Apache's Hadoop project. It uses JSON for defining data types and protocols, and serializes data in a compact binary format.

  7. Cascading (software) - Wikipedia

    en.wikipedia.org/wiki/Cascading_(software)

    Cascading is a software abstraction layer for Apache Hadoop and Apache Flink. Cascading is used to create and execute complex data processing workflows on a Hadoop cluster using any JVM-based language (Java, JRuby, Clojure, etc.), hiding the underlying complexity of MapReduce jobs. It is open source and available under the Apache License.

  8. Apache Parquet - Wikipedia

    en.wikipedia.org/wiki/Apache_Parquet

    Apache Parquet is a free and open-source column-oriented data storage format in the Apache Hadoop ecosystem. It is similar to RCFile and ORC, the other columnar-storage file formats in Hadoop, and is compatible with most of the data processing frameworks around Hadoop.

  9. Hue (software) - Wikipedia

    en.wikipedia.org/wiki/Hue_(Software)

    Hue is an open-source SQL Assistant for querying Databases & Data Warehouses and collaborating. Its goal is to make self service data querying more widespread in organizations.