Search Hadoop -
Colin Patrick McCabe
Arun C Murthy
Tsz Wo Nicholas Sze
Brahma Reddy Battula
Vinod Kumar Vavilapalli
Robert Joseph Evans
last 7 days (0)
last 30 days (0)
last 90 days (0)
last 6 months (0)
last 9 months (59)
Solr & Elasticsearch trainings in New York & San Francisco
San Francisco - Oct 4-6
New York - Oct 10-12
San Francisco - Oct 4-7
New York - Oct 10-12
and all its subprojects:
newest on top
oldest on top
. Results from
Loading phrases to help you
refine your search...
[HADOOP-11794] distcp can copy blocks in parallel
...The minimum unit of work for a distcp task is a file. We have files that are greater than 1 TB with a block size of 1 GB. If we use distcp to copy these files, the tasks either take a ...
, 2016-01-29, 22:12
[HADOOP-2655] Copy on write for data and metadata files in the presence of snapshots
...If a DFS Client wants to append data to an existing file (appends, HADOOP-1700) and a snapshot is present, the Datanoed has to implement some form of a copy-on-write for writes to data and m...
, 2015-09-08, 23:43
[HADOOP-3834] Checkin the design document for HDFS appends into source control repository
...The design document for HDFS needs to be converted into forrest and checked into repository....
, 2014-07-18, 19:22
[HADOOP-4565] MultiFileInputSplit can use data locality information to create splits
...The MultiFileInputFormat takes a set of paths and creates splits based on file sizes. Each splits contains a few files an each split are roughly equal in size. It would be efficient if we ca...
, 2014-07-17, 20:31
[HADOOP-6149] FileStatus can support a fileid per path
...FileStatus should expose a id that uniquely identifies a file. This helps in developing applications that work correctly even when files are moved from one directory to another....
, 2013-04-19, 23:31
[HADOOP-923] DFS Scalability: datanode heartbeat timeouts cause cascading timeouts of other datanodes
...The datanode sends a heartbeat to the namenode every 3 seconds. The namenode processes the heartbeat and sends a list of block-to-be-replicated and blocks-to-be-deleted as part of the ...
, 2012-09-18, 18:18
[HADOOP-7460] Support for pluggable Trash policies
...It would be beneficial to make the Trash policy pluggable. One primary use-case for this is to archive files (in some remote store) when they get removed by Trash emptier....
, 2012-07-15, 22:18
[HADOOP-2330] Preallocate transaction log to improve namenode transaction logging performance
...In the current implementation, the transaction log is opened in "append" mode and every new transaction is written to the end of the log. This means that new blocks get allocated to the edit...
, 2012-06-15, 22:50
[HADOOP-970] Reduce CPU usage of hadoop ipc package
...There are a couple of optimizations that could be done to reduce CPU consumption.1. The method Server.cleanupConnections() could be invoked less often.2. The method Server.cleanupConnections...
, 2011-07-16, 18:26
[HADOOP-1117] DFS Scalability: When the namenode is restarted it consumes 80% CPU
...When the namenode is restarted, the datanodes register and each block is inserted into neededReplication. When the namenode exists, safemode it sees starts processing neededReplication. It p...
, 2011-05-16, 21:32
Apache Lucene, Apache Solr and all other Apache Software Foundation projects and their respective logos are trademarks of the Apache Software Foundation.
Elasticsearch, Kibana, Logstash, and Beats are trademarks of Elasticsearch BV, registered in the U.S. and in other countries. This site and Sematext Group is in no way affiliated with Elasticsearch BV.
Service operated by