本文整理了Java中org.apache.hadoop.hdfs.protocol.Block.setBlockId()
方法的一些代码示例,展示了Block.setBlockId()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。Block.setBlockId()
方法的具体详情如下:
包路径:org.apache.hadoop.hdfs.protocol.Block
类名称:Block
方法名:setBlockId
暂无
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs
/**
* @param b A block object whose id is set to the starting point for check
* @return true if any ID in the range
* {id, id+HdfsConstants.MAX_BLOCKS_IN_GROUP} is pointed-to by a stored
* block.
*/
private boolean hasValidBlockInRange(Block b) {
final long id = b.getBlockId();
for (int i = 0; i < MAX_BLOCKS_IN_GROUP; i++) {
b.setBlockId(id + i);
if (blockManager.getStoredBlock(b) != null) {
return true;
}
}
return false;
}
}
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs
/**
* Identify the block stored in the given datanode storage. Note that
* the returned block has the same block Id with the one seen/reported by the
* DataNode.
*/
Block getBlockOnStorage(DatanodeStorageInfo storage) {
int index = getStorageBlockIndex(storage);
if (index < 0) {
return null;
} else {
Block block = new Block(this);
block.setBlockId(this.getBlockId() + index);
return block;
}
}
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs
@Override // NumberGenerator
public long nextValue() {
Block b = new Block(super.nextValue());
// There may be an occasional conflict with randomly generated
// block IDs. Skip over the conflicts.
while(isValidBlock(b)) {
b.setBlockId(super.nextValue());
}
if (b.getBlockId() < 0) {
throw new IllegalStateException("All positive block IDs are used, " +
"wrapping to negative IDs, " +
"which might conflict with erasure coded block groups.");
}
return b.getBlockId();
}
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs
public DBlock getInternalBlock(StorageGroup storage) {
int idxInLocs = locations.indexOf(storage);
if (idxInLocs == -1) {
return null;
}
byte idxInGroup = indices[idxInLocs];
long blkId = getBlock().getBlockId() + idxInGroup;
long numBytes = getInternalBlockLength(getNumBytes(), cellSize,
dataBlockNum, idxInGroup);
Block blk = new Block(getBlock());
blk.setBlockId(blkId);
blk.setNumBytes(numBytes);
DBlock dblk = new DBlock(blk);
dblk.addLocation(storage);
return dblk;
}
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs
@Override // NumberGenerator
public long nextValue() {
skipTo((getCurrentValue() & ~BLOCK_GROUP_INDEX_MASK) + MAX_BLOCKS_IN_GROUP);
// Make sure there's no conflict with existing random block IDs
final Block b = new Block(getCurrentValue());
while (hasValidBlockInRange(b)) {
skipTo(getCurrentValue() + MAX_BLOCKS_IN_GROUP);
b.setBlockId(getCurrentValue());
}
if (b.getBlockId() >= 0) {
throw new IllegalStateException("All negative block group IDs are used, "
+ "growing into positive IDs, "
+ "which might conflict with non-erasure coded blocks.");
}
return getCurrentValue();
}
代码示例来源:origin: io.prestosql.hadoop/hadoop-apache
public void setBlockId(final long bid) {
block.setBlockId(bid);
}
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs-client
public void setBlockId(final long bid) {
block.setBlockId(bid);
}
代码示例来源:origin: ch.cern.hadoop/hadoop-hdfs
public void setBlockId(final long bid) {
block.setBlockId(bid);
}
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs
Block blk = new Block(ExtendedBlock.getLocalBlock(extBlock));
long blkId = blk.getBlockId() + blkIndex;
blk.setBlockId(blkId);
blk.setNumBytes(numBytes);
BlockMovingInfo blkMovingInfo = new BlockMovingInfo(blk, sourceNode,
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs
Block bi = new Block(storedBlock);
if (storedBlock.isStriped()) {
bi.setBlockId(bi.getBlockId() + i);
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs
corrupted.setBlockId(b.getStored().getBlockId());
代码示例来源:origin: com.facebook.hadoop/hadoop-core
/**
* Allocate a block at the given pending filename
*
* @param src path to the file
* @param inodes INode representing each of the components of src.
* <code>inodes[inodes.length-1]</code> is the INode for the file.
*/
private Block allocateBlock(String src, INode[] inodes) throws IOException {
Block b = new Block(FSNamesystem.randBlockId.nextLong(), 0, 0);
while (isValidBlock(b)) {
b.setBlockId(FSNamesystem.randBlockId.nextLong());
}
b.setGenerationStamp(getGenerationStamp());
b = dir.addBlock(src, inodes, b);
return b;
}
代码示例来源:origin: ch.cern.hadoop/hadoop-hdfs
@Override // NumberGenerator
public long nextValue() {
Block b = new Block(super.nextValue());
// There may be an occasional conflict with randomly generated
// block IDs. Skip over the conflicts.
while(isValidBlock(b)) {
b.setBlockId(super.nextValue());
}
return b.getBlockId();
}
代码示例来源:origin: com.facebook.hadoop/hadoop-core
public static void getBlockInfo(Block b, long[] blockArray, int index){
b.setBlockId(blockArray[index2BlockId(index)]);
b.setNumBytes(blockArray[index2BlockLen(index)]);
b.setGenerationStamp(blockArray[index2BlockGenStamp(index)]);
}
代码示例来源:origin: io.prestosql.hadoop/hadoop-apache
@Override // NumberGenerator
public long nextValue() {
Block b = new Block(super.nextValue());
// There may be an occasional conflict with randomly generated
// block IDs. Skip over the conflicts.
while(isValidBlock(b)) {
b.setBlockId(super.nextValue());
}
return b.getBlockId();
}
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs-test
blk.setBlockId(blkid-1);
DataTransferProtocol.Sender.opReadBlock(sendOut, blk, 0L, fileLen, "cl",
BlockTokenSecretManager.DUMMY_TOKEN);
blk.setBlockId(blkid);
DataTransferProtocol.Sender.opReadBlock(sendOut, blk, -1L, fileLen, "cl",
BlockTokenSecretManager.DUMMY_TOKEN);
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs-test
b.setBlockId(0);
assertNull(map.get(b));
代码示例来源:origin: org.apache.hadoop/hadoop-hdfs-test
@Test
public void testRemove() {
// Test 1: null argument throws invalid argument exception
try {
map.remove(null);
fail("Expected exception not thrown");
} catch (IllegalArgumentException expected) { }
// Test 2: remove failure - generation stamp mismatch
Block b = new Block(block);
b.setGenerationStamp(0);
assertNull(map.remove(b));
// Test 3: remove failure - blockID mismatch
b.setGenerationStamp(block.getGenerationStamp());
b.setBlockId(0);
assertNull(map.remove(b));
// Test 4: remove success
assertNotNull(map.remove(block));
// Test 5: remove failure - invalid blockID
assertNull(map.remove(0));
// Test 6: remove success
map.add(new FinalizedReplica(block, null, null));
assertNotNull(map.remove(block.getBlockId()));
}
}
代码示例来源:origin: ch.cern.hadoop/hadoop-hdfs
b.setBlockId(0);
assertNull(map.get(bpid, b));
代码示例来源:origin: ch.cern.hadoop/hadoop-hdfs
@Test
public void testRemove() {
// Test 1: null argument throws invalid argument exception
try {
map.remove(bpid, null);
fail("Expected exception not thrown");
} catch (IllegalArgumentException expected) { }
// Test 2: remove failure - generation stamp mismatch
Block b = new Block(block);
b.setGenerationStamp(0);
assertNull(map.remove(bpid, b));
// Test 3: remove failure - blockID mismatch
b.setGenerationStamp(block.getGenerationStamp());
b.setBlockId(0);
assertNull(map.remove(bpid, b));
// Test 4: remove success
assertNotNull(map.remove(bpid, block));
// Test 5: remove failure - invalid blockID
assertNull(map.remove(bpid, 0));
// Test 6: remove success
map.add(bpid, new FinalizedReplica(block, null, null));
assertNotNull(map.remove(bpid, block.getBlockId()));
}
}
内容来源于网络,如有侵权,请联系作者删除!