[RFR] templates://hadoop/{hadoop-namenoded.templates}

Justin B Rye jbr at edlug.org.uk
Tue Mar 30 17:01:19 UTC 2010


Justin B Rye wrote:
> comments but no actual patch attached.

Second thoughts and patch.
 
>>  Template: hadoop-namenoded/format
> [...]
>    The namenode manages the Hadoop Distributed File System (HDFS). Like a
>    normal file system, it needs to be formatted before use; otherwise
>    the namenode daemon will not start.

For consistency with the package descriptions maybe it should be
 jobtrackerd	= the Job Tracker (daemon)
 tasktrackerd	= the Task Tracker (daemon)
 namenoded	= the Name Node (daemon)
so:

 _Description: Should namenoded's file system be formatted?
  The Name Node daemon manages the Hadoop Distributed File System (HDFS).
  Like a normal file system, it needs to be formatted prior to first use.
  If the HDFS file system is not formatted, the Name Node will fail to
  start.

(Or for extra terseness we could push the Description in the
direction of "Format HDFS?")

The package description boilerplate:
> I'm not sure they should all carry all three boilerplate paragraphs; maybe
> since hadoop-bin is a common dependency it makes sense for it to carry the
> "long version".

I was going to take out the third paragraph, but then I noticed it
wasn't in most of the package descriptions anyway.

>>  Package: libhadoop-index-java
>    Description: data-intensive clustering framework - Lucene index support
[...] 
>> + Hadoop implements MapReduce, using the Hadoop Distributed File System (HDFS).
>> + MapReduce divides applications into many small blocks of work. HDFS creates
>> + multiple replicas of data blocks for reliability, placing them on compute
>> + nodes around the cluster. MapReduce can then process the data where it is
>> + located.
>>   .
>>   This contrib package provides a utility to build or update an index
>>   using Map/Reduce.
>>   .
>>   A distributed "index" is partitioned into "shards". Each shard corresponds
>>   to a Lucene instance. org.apache.hadoop.contrib.index.main.UpdateIndex
>>   contains the main() method which uses a Map/Reduce job to analyze documents
>>   and update Lucene instances in parallel.

Drop the first explanation of MapReduce and the misleading phrase
"contrib package"; merge the remainder into one package-specific
paragraph, and make it clearer exactly what it's describing.

     The org.apache.hadoop.contrib.index.main.UpdateIndex library provides
     support for managing an index using MapReduce. A distributed "index" is
     partitioned into "shards", each corresponding to a Lucene instance.
     This library's main() method uses a MapReduce job to analyze documents
     and update Lucene instances in parallel.

(Is it canonically MapReduce, not Map/Reduce?)

>> Package: hadoop-bin

Hang on, I hadn't noticed this contains /usr/bin/hadoop (which is an
executable, but not a binary one; it's a shell wrapperscript).  So:

     This package provides the hadoop command line interface. See the hadoop-.*d
     packages for the Hadoop daemons.

>> Package: hadoop-daemons-common
> [...]
>> + This package prepares some common things for all hadoop daemon packages:
>>    * creates the user hadoop
>>    * creates data and log directories owned by the hadoop user
>>    * manages the update-alternatives mechanism for hadoop configuration

Rephrase as:
     This package provides infrastructure for the Hadoop daemon packages,
     creating the hadoop user (with data and log directories) and maintaining
     the update-alternatives mechanism for Hadoop configuration.

-- 
JBR	with qualifications in linguistics, experience as a Debian
	sysadmin, and probably no clue about this particular package
-------------- next part --------------
--- ../hadoop-0.20.2+dfsg1.pristine/debian/hadoop-namenoded.templates	2010-03-05 17:54:15.000000000 +0000
+++ debian/hadoop-namenoded.templates	2010-03-30 17:44:29.000000000 +0100
@@ -1,17 +1,16 @@
 Template: hadoop-namenoded/format
 Type: boolean
 Default: false
-_Description: Should the namenode's filesystem be formatted now?
- The namenode manages the Hadoop Distributed FileSystem (HDFS). Like a
- normal filesystem, it needs to be formatted prior to first use. If the
- HDFS filesystem is not formatted, the namenode daemon will fail to
+_Description: Should namenoded's file system be formatted?
+ The Name Node daemon manages the Hadoop Distributed File System (HDFS).
+ Like a normal file system, it needs to be formatted prior to first use.
+ If the HDFS file system is not formatted, the Name Node will fail to
  start.
  .
- This operation does not affect the "normal" filesystem on this
- computer. If you're using HDFS for the first time and don't have data
- from previous installations on this computer, it should be save to
- proceed with yes.
+ This operation does not affect other file systems on this
+ computer. You can safely choose to format the file system if you're
+ using HDFS for the first time and don't have data from previous
+ installations on this computer.
  .
- You can later on format the filesystem yourself with
- . 
- su -c"hadoop namenode -format" hadoop
+ If you choose not to format the file system right now, you can do it
+ later by executing "hadoop namenode -format" as the user "hadoop".
--- ../hadoop-0.20.2+dfsg1.pristine/debian/control	2010-03-04 10:57:15.000000000 +0000
+++ debian/control	2010-03-30 17:49:56.000000000 +0100
@@ -29,7 +29,7 @@
 
 Package: libhadoop-java
 Architecture: all
-Depends: ${misc:Depends}, 
+Depends: ${misc:Depends},
  libcommons-cli-java,
  libcommons-codec-java,
  libcommons-el-java,
@@ -44,20 +44,46 @@
  libslf4j-java,
  libxmlenc-java
 Suggests: libhsqldb-java
-Description: software platform for processing vast amounts of data
- This package contains the core java libraries.
+Description: data-intensive clustering framework - Java libraries
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
+ This package contains the core Java libraries.
 
 Package: libhadoop-index-java
 Architecture: all
 Depends: ${misc:Depends}, libhadoop-java (= ${binary:Version}),
  liblucene2-java
-Description: Hadoop contrib to create lucene indexes
- This contrib package provides a utility to build or update an index
- using Map/Reduce.
- .
- A distributed "index" is partitioned into "shards". Each shard corresponds
- to a Lucene instance. org.apache.hadoop.contrib.index.main.UpdateIndex
- contains the main() method which uses a Map/Reduce job to analyze documents
+Description: data-intensive clustering framework - Lucene index support
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
+ The org.apache.hadoop.contrib.index.main.UpdateIndex library provides
+ support for managing an index using MapReduce. A distributed "index" is
+ partitioned into "shards", each corresponding to a Lucene instance.
+ This library's main() method uses a MapReduce job to analyze documents
  and update Lucene instances in parallel.
 
 Package: hadoop-bin
@@ -65,9 +91,9 @@
 Architecture: all
 Depends: ${misc:Depends}, libhadoop-java (= ${binary:Version}),
  default-jre-headless | java6-runtime-headless
-Description: software platform for processing vast amounts of data
- Hadoop is a software platform that lets one easily write and
- run applications that process vast amounts of data.
+Description: data-intensive clustering framework - tools
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
  .
  Here's what makes Hadoop especially useful:
   * Scalable: Hadoop can reliably store and process petabytes.
@@ -86,33 +112,73 @@
  nodes around the cluster. MapReduce can then process the data where it is
  located.
  .
- This package contains the hadoop shell interface. See the packages hadoop-.*d
- for the hadoop daemons.
+ This package provides the hadoop command line interface. See the hadoop-.*d
+ packages for the Hadoop daemons.
 
 Package: hadoop-daemons-common
 Section: misc
 Architecture: all
 Depends: ${misc:Depends}, hadoop-bin (= ${binary:Version}), daemon, adduser,
  lsb-base (>= 3.2-14)
-Description: Creates user and directories for hadoop daemons
- Prepares some common things for all hadoop daemon packages:
-  * creates the user hadoop
-  * creates data and log directories owned by the hadoop user
-  * manages the update-alternatives mechanism for hadoop configuration
-  * brings in the common dependencies
+Description: data-intensive clustering framework - common files
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
+ This package provides infrastructure for the Hadoop daemon packages,
+ creating the hadoop user (with data and log directories) and maintaining
+ the update-alternatives mechanism for hadoop configuration.
 
 Package: libhadoop-java-doc
 Section: doc
 Architecture: all
 Depends: ${misc:Depends}, libhadoop-java (= ${binary:Version})
-Description: Contains the javadoc for hadoop
- contains the api documentation of hadoop
+Description: data-intensive clustering framework - Java documentation
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
+ This package provides the API documentation of Hadoop.
 
 Package: hadoop-tasktrackerd
 Section: misc
 Architecture: all
 Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
-Description: Task Tracker for Hadoop
+Description: data-intensive clustering framework - Task Tracker
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
  The Task Tracker is the Hadoop service that accepts MapReduce tasks and
  computes results. Each node in a Hadoop cluster that should be doing
  computation should run a Task Tracker.
@@ -121,34 +187,90 @@
 Section: misc
 Architecture: all
 Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
-Description: Job Tracker for Hadoop
- The jobtracker is a central service which is responsible for managing
- the tasktracker services running on all nodes in a Hadoop Cluster.
- The jobtracker allocates work to the tasktracker nearest to the data
+Description: data-intensive clustering framework - Job Tracker
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
+ The Job Tracker is a central service which is responsible for managing
+ the Task Tracker services running on all nodes in an Hadoop Cluster.
+ The Job Tracker allocates work to the Task Tracker nearest to the data
  with an available work slot.
 
 Package: hadoop-namenoded
 Section: misc
 Architecture: all
 Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
-Description: Name Node for Hadoop
- The Hadoop Distributed Filesystem (HDFS) requires one unique server, the
- namenode, which manages the block locations of files on the filesystem.
+Description: data-intensive clustering framework - Name Node
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
+ The Hadoop Distributed File System (HDFS) requires one unique server, the
+ Name Node, which manages the block locations of files on the file system.
 
 Package: hadoop-secondarynamenoded
 Section: misc
 Architecture: all
 Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
-Description: Secondary Name Node for Hadoop
- The Secondary Name Node is responsible for checkpointing file system images.
- It is _not_ a failover pair for the namenode, and may safely be run on the
- same machine.
+Description: data-intensive clustering framework - secondary Name Node
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
+ The secondary Name Node is responsible for checkpointing file system images.
+ It is _not_ a failover partner for the name node, and may safely be run on
+ the same machine.
 
 Package: hadoop-datanoded
 Section: misc
 Architecture: all
 Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
-Description: Data Node for Hadoop
+Description: data-intensive clustering framework - Data Node
+ Hadoop is a software platform for writing and running applications
+ that process vast amounts of data on a distributed file system.
+ .
+ Here's what makes Hadoop especially useful:
+  * Scalable: Hadoop can reliably store and process petabytes.
+  * Economical: It distributes the data and processing across clusters
+                of commonly available computers. These clusters can number
+                into the thousands of nodes.
+  * Efficient: By distributing the data, Hadoop can process it in parallel
+               on the nodes where the data is located. This makes it
+               extremely rapid.
+  * Reliable: Hadoop automatically maintains multiple copies of data and
+              automatically redeploys computing tasks based on failures.
+ .
  The Data Nodes in the Hadoop Cluster are responsible for serving up
- blocks of data over the network to Hadoop Distributed Filesystem
+ blocks of data over the network to Hadoop Distributed File System
  (HDFS) clients.
-------------- next part --------------
Template: hadoop-namenoded/format
Type: boolean
Default: false
_Description: Should namenoded's file system be formatted?
 The Name Node daemon manages the Hadoop Distributed File System (HDFS).
 Like a normal file system, it needs to be formatted prior to first use.
 If the HDFS file system is not formatted, the Name Node will fail to
 start.
 .
 This operation does not affect other file systems on this
 computer. You can safely choose to format the file system if you're
 using HDFS for the first time and don't have data from previous
 installations on this computer.
 .
 If you choose not to format the file system right now, you can do it
 later by executing "hadoop namenode -format" as the user "hadoop".
-------------- next part --------------
Source: hadoop
Section: java
Priority: optional
Maintainer: Debian Java Maintainers <pkg-java-maintainers at lists.alioth.debian.org>
Uploaders: Thomas Koch <thomas.koch at ymc.ch>
Homepage: http://hadoop.apache.org
Vcs-Browser: http://git.debian.org/?p=pkg-java/hadoop.git
Vcs-Git: git://git.debian.org/pkg-java/hadoop.git
Standards-Version: 3.8.4
Build-Depends: debhelper (>= 7.4.11), default-jdk, ant (>= 1.6.0), javahelper (>= 0.28),
 po-debconf,
 libcommons-cli-java,
 libcommons-codec-java,
 libcommons-el-java,
 libcommons-httpclient-java,
 libcommons-io-java,
 libcommons-logging-java,
 libcommons-net-java,
 libtomcat6-java,
 libjetty-java (>>6),
 libservlet2.5-java,
 liblog4j1.2-java,
 libslf4j-java,
 libxmlenc-java,
 liblucene2-java,
 libhsqldb-java,
 ant-optional,
 javacc

Package: libhadoop-java
Architecture: all
Depends: ${misc:Depends},
 libcommons-cli-java,
 libcommons-codec-java,
 libcommons-el-java,
 libcommons-httpclient-java,
 libcommons-io-java,
 libcommons-logging-java,
 libcommons-net-java,
 libtomcat6-java,
 libjetty-java (>>6),
 libservlet2.5-java,
 liblog4j1.2-java,
 libslf4j-java,
 libxmlenc-java
Suggests: libhsqldb-java
Description: data-intensive clustering framework - Java libraries
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 This package contains the core Java libraries.

Package: libhadoop-index-java
Architecture: all
Depends: ${misc:Depends}, libhadoop-java (= ${binary:Version}),
 liblucene2-java
Description: data-intensive clustering framework - Lucene index support
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 The org.apache.hadoop.contrib.index.main.UpdateIndex library provides
 support for managing an index using MapReduce. A distributed "index" is
 partitioned into "shards", each corresponding to a Lucene instance.
 This library's main() method uses a MapReduce job to analyze documents
 and update Lucene instances in parallel.

Package: hadoop-bin
Section: misc
Architecture: all
Depends: ${misc:Depends}, libhadoop-java (= ${binary:Version}),
 default-jre-headless | java6-runtime-headless
Description: data-intensive clustering framework - tools
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 Hadoop implements MapReduce, using the Hadoop Distributed File System (HDFS).
 MapReduce divides applications into many small blocks of work. HDFS creates
 multiple replicas of data blocks for reliability, placing them on compute
 nodes around the cluster. MapReduce can then process the data where it is
 located.
 .
 This package provides the hadoop command line interface. See the hadoop-.*d
 packages for the Hadoop daemons.

Package: hadoop-daemons-common
Section: misc
Architecture: all
Depends: ${misc:Depends}, hadoop-bin (= ${binary:Version}), daemon, adduser,
 lsb-base (>= 3.2-14)
Description: data-intensive clustering framework - common files
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 This package provides infrastructure for the Hadoop daemon packages,
 creating the hadoop user (with data and log directories) and maintaining
 the update-alternatives mechanism for hadoop configuration.

Package: libhadoop-java-doc
Section: doc
Architecture: all
Depends: ${misc:Depends}, libhadoop-java (= ${binary:Version})
Description: data-intensive clustering framework - Java documentation
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 This package provides the API documentation of Hadoop.

Package: hadoop-tasktrackerd
Section: misc
Architecture: all
Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
Description: data-intensive clustering framework - Task Tracker
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 The Task Tracker is the Hadoop service that accepts MapReduce tasks and
 computes results. Each node in a Hadoop cluster that should be doing
 computation should run a Task Tracker.

Package: hadoop-jobtrackerd
Section: misc
Architecture: all
Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
Description: data-intensive clustering framework - Job Tracker
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 The Job Tracker is a central service which is responsible for managing
 the Task Tracker services running on all nodes in an Hadoop Cluster.
 The Job Tracker allocates work to the Task Tracker nearest to the data
 with an available work slot.

Package: hadoop-namenoded
Section: misc
Architecture: all
Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
Description: data-intensive clustering framework - Name Node
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 The Hadoop Distributed File System (HDFS) requires one unique server, the
 Name Node, which manages the block locations of files on the file system.

Package: hadoop-secondarynamenoded
Section: misc
Architecture: all
Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
Description: data-intensive clustering framework - secondary Name Node
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 The secondary Name Node is responsible for checkpointing file system images.
 It is _not_ a failover partner for the name node, and may safely be run on
 the same machine.

Package: hadoop-datanoded
Section: misc
Architecture: all
Depends: ${misc:Depends}, hadoop-daemons-common (= ${binary:Version})
Description: data-intensive clustering framework - Data Node
 Hadoop is a software platform for writing and running applications
 that process vast amounts of data on a distributed file system.
 .
 Here's what makes Hadoop especially useful:
  * Scalable: Hadoop can reliably store and process petabytes.
  * Economical: It distributes the data and processing across clusters
                of commonly available computers. These clusters can number
                into the thousands of nodes.
  * Efficient: By distributing the data, Hadoop can process it in parallel
               on the nodes where the data is located. This makes it
               extremely rapid.
  * Reliable: Hadoop automatically maintains multiple copies of data and
              automatically redeploys computing tasks based on failures.
 .
 The Data Nodes in the Hadoop Cluster are responsible for serving up
 blocks of data over the network to Hadoop Distributed File System
 (HDFS) clients.


More information about the pkg-java-maintainers mailing list