2009년 7월 23일 목요일

What Is Hadoop?

다음은 apache.org 에 소개된 Hadoop에 관한 내용이다.


The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing. Hadoop includes these subprojects:

  • Hadoop Common: The common utilities that support the other Hadoop subprojects.
  • Avro: A data serialization system that provides dynamic integration with scripting languages.
  • Chukwa: A data collection system for managing large distributed systems.
  • HBase: A scalable, distributed database that supports structured data storage for large tables.
  • HDFS: A distributed file system that provides high throughput access to application data.
  • Hive: A data warehouse infrastructure that provides data summarization and ad hoc querying.
  • MapReduce: A software framework for distributed processing of large data sets on compute clusters.
  • Pig: A high-level data-flow language and execution framework for parallel computation.
  • ZooKeeper: A high-performance coordination service for distributed applications.

Hadoop Common subprojectHadoop Core 로 명칭이 변경되었다.

2009년 7월 20일 월요일

textcube 입성

드뎌 textcube 에 입성하였다.

글 쓰는 첫날 오류가 나네... ^^

점점 좋아지는 TextCube를 기대하며...
2009년 7월 20일