Java 类org.apache.hadoop.hdfs.server.datanode.metrics.DataNodeMetrics 实例源码

项目:hadoop    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:aliyun-oss-hadoop-fs    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:big-c    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:hadoop-2.6.0-cdh5.4.3    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:hadoop-EAR    文件:DataNode.java   
/**
 * Initialize global settings for DN
 */
protected void initGlobalSetting(Configuration conf,
    AbstractList<File> dataDirs) throws IOException {
  this.dataDirs = dataDirs;
  this.conf = conf;
  storage = new DataStorage(this);

  // global DN settings
  initConfig(conf);
  registerMXBean();
  initDataXceiver(conf);
  startInfoServer(conf);
  initIpcServer(conf);

  myMetrics = new DataNodeMetrics(conf, storage.getStorageID());
  setCountingLoggers(myMetrics);
  threadLivenessReporter = new DatanodeThreadLivenessReporter(conf.getLong(
      "dfs.datanode.thread.liveness.threshold", 240 * 1000),
      myMetrics.threadActiveness);
}
项目:hadoop-plus    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:FlexMap    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:hadoop-TCP    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:hardfs    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:hadoop-on-lustre2    文件:TestBPOfferService.java   
@Before
public void setupMocks() throws Exception {
  mockNN1 = setupNNMock(0);
  mockNN2 = setupNNMock(1);

  // Set up a mock DN with the bare-bones configuration
  // objects, etc.
  mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(new File(TEST_BUILD_DATA, "dfs"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn"))
  .when(mockDn).getMetrics();

  // Set up a simulated dataset with our fake BP
  mockFSDataset = Mockito.spy(new SimulatedFSDataset(null, conf));
  mockFSDataset.addBlockPool(FAKE_BPID, conf);

  // Wire the dataset to the DN.
  Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
}
项目:cumulus    文件:TestDataNodeMetrics.java   
public void testDataNodeMetrics() throws Exception {
  Configuration conf = new HdfsConfiguration();
  conf.setBoolean(SimulatedFSDataset.CONFIG_PROPERTY_SIMULATED, true);
  MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf).build();
  try {
    FileSystem fs = cluster.getFileSystem();
    final long LONG_FILE_LEN = Integer.MAX_VALUE+1L; 
    DFSTestUtil.createFile(fs, new Path("/tmp.txt"),
        LONG_FILE_LEN, (short)1, 1L);
    List<DataNode> datanodes = cluster.getDataNodes();
    assertEquals(datanodes.size(), 1);
    DataNode datanode = datanodes.get(0);
    DataNodeMetrics metrics = datanode.getMetrics();
    assertEquals(LONG_FILE_LEN, metrics.bytesWritten.getCurrentIntervalValue());
  } finally {
    if (cluster != null) {cluster.shutdown();}
  }
}
项目:RDFS    文件:DataNode.java   
/**
 * Initialize global settings for DN
 */
protected void initGlobalSetting(Configuration conf,
    AbstractList<File> dataDirs) throws IOException {
  this.dataDirs = dataDirs;
  this.conf = conf;
  storage = new DataStorage(this);

  // global DN settings
  initConfig(conf);
  registerMXBean();
  initDataXceiver(conf);
  startInfoServer(conf);
  initIpcServer(conf);

  myMetrics = new DataNodeMetrics(conf, storage.getStorageID());
}
项目:hadoop-0.20    文件:TestDataNodeMetrics.java   
public void testDataNodeMetrics() throws Exception {
  Configuration conf = new Configuration();
  conf.setBoolean(SimulatedFSDataset.CONFIG_PROPERTY_SIMULATED, true);
  MiniDFSCluster cluster = new MiniDFSCluster(conf, 1, true, null);
  try {
    FileSystem fs = cluster.getFileSystem();
    final long LONG_FILE_LEN = Integer.MAX_VALUE+1L; 
    DFSTestUtil.createFile(fs, new Path("/tmp.txt"),
        LONG_FILE_LEN, (short)1, 1L);
    List<DataNode> datanodes = cluster.getDataNodes();
    assertEquals(datanodes.size(), 1);
    DataNode datanode = datanodes.get(0);
    DataNodeMetrics metrics = datanode.getMetrics();
    assertEquals(LONG_FILE_LEN, metrics.bytesWritten.getCurrentIntervalValue());
  } finally {
    if (cluster != null) {cluster.shutdown();}
  }
}
项目:hadoop-gpu    文件:TestDataNodeMetrics.java   
public void testDataNodeMetrics() throws Exception {
  Configuration conf = new Configuration();
  conf.setBoolean(SimulatedFSDataset.CONFIG_PROPERTY_SIMULATED, true);
  MiniDFSCluster cluster = new MiniDFSCluster(conf, 1, true, null);
  try {
    FileSystem fs = cluster.getFileSystem();
    final long LONG_FILE_LEN = Integer.MAX_VALUE+1L; 
    DFSTestUtil.createFile(fs, new Path("/tmp.txt"),
        LONG_FILE_LEN, (short)1, 1L);
    List<DataNode> datanodes = cluster.getDataNodes();
    assertEquals(datanodes.size(), 1);
    DataNode datanode = datanodes.get(0);
    DataNodeMetrics metrics = datanode.getMetrics();
    assertEquals(LONG_FILE_LEN, metrics.bytesWritten.getCurrentIntervalValue());
  } finally {
    if (cluster != null) {cluster.shutdown();}
  }
}
项目:hadoop    文件:TestBPOfferService.java   
/**
 * Test datanode block pool initialization error handling.
 * Failure in initializing a block pool should not cause NPE.
 */
@Test
public void testBPInitErrorHandling() throws Exception {
  final DataNode mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(
    new File(TEST_BUILD_DATA, "testBPInitErrorHandling"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn")).
    when(mockDn).getMetrics();
  final AtomicInteger count = new AtomicInteger();
  Mockito.doAnswer(new Answer<Void>() {
    @Override
    public Void answer(InvocationOnMock invocation) throws Throwable {
      if (count.getAndIncrement() == 0) {
        throw new IOException("faked initBlockPool exception");
      }
      // The initBlockPool is called again. Now mock init is done.
      Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
      return null;
    }
  }).when(mockDn).initBlockPool(Mockito.any(BPOfferService.class));
  BPOfferService bpos = setupBPOSForNNs(mockDn, mockNN1, mockNN2);
  List<BPServiceActor> actors = bpos.getBPServiceActors();
  assertEquals(2, actors.size());
  bpos.start();
  try {
    waitForInitialization(bpos);
    // even if one of the actor initialization fails, the other one will be
    // finish block report.
    waitForBlockReport(mockNN1, mockNN2);
  } finally {
    bpos.stop();
  }
}
项目:aliyun-oss-hadoop-fs    文件:TestBPOfferService.java   
/**
 * Test datanode block pool initialization error handling.
 * Failure in initializing a block pool should not cause NPE.
 */
@Test
public void testBPInitErrorHandling() throws Exception {
  final DataNode mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(
    new File(TEST_BUILD_DATA, "testBPInitErrorHandling"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn")).
    when(mockDn).getMetrics();
  final AtomicInteger count = new AtomicInteger();
  Mockito.doAnswer(new Answer<Void>() {
    @Override
    public Void answer(InvocationOnMock invocation) throws Throwable {
      if (count.getAndIncrement() == 0) {
        throw new IOException("faked initBlockPool exception");
      }
      // The initBlockPool is called again. Now mock init is done.
      Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
      return null;
    }
  }).when(mockDn).initBlockPool(Mockito.any(BPOfferService.class));
  BPOfferService bpos = setupBPOSForNNs(mockDn, mockNN1, mockNN2);
  List<BPServiceActor> actors = bpos.getBPServiceActors();
  assertEquals(2, actors.size());
  bpos.start();
  try {
    waitForInitialization(bpos);
    // even if one of the actor initialization fails, the other one will be
    // finish block report.
    waitForBlockReport(mockNN1, mockNN2);
  } finally {
    bpos.stop();
    bpos.join();
  }
}
项目:aliyun-oss-hadoop-fs    文件:TestDataXceiverLazyPersistHint.java   
private static DataNode getMockDn(NonLocalLazyPersist nonLocalLazyPersist) {
  Configuration conf = new HdfsConfiguration();
  conf.setBoolean(
      DFS_DATANODE_NON_LOCAL_LAZY_PERSIST,
      nonLocalLazyPersist == NonLocalLazyPersist.ALLOWED);
  DNConf dnConf = new DNConf(conf);
  DataNodeMetrics mockMetrics = mock(DataNodeMetrics.class);
  DataNode mockDn = mock(DataNode.class);
  when(mockDn.getDnConf()).thenReturn(dnConf);
  when(mockDn.getConf()).thenReturn(conf);
  when(mockDn.getMetrics()).thenReturn(mockMetrics);
  return mockDn;
}
项目:big-c    文件:TestBPOfferService.java   
/**
 * Test datanode block pool initialization error handling.
 * Failure in initializing a block pool should not cause NPE.
 */
@Test
public void testBPInitErrorHandling() throws Exception {
  final DataNode mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(
    new File(TEST_BUILD_DATA, "testBPInitErrorHandling"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn")).
    when(mockDn).getMetrics();
  final AtomicInteger count = new AtomicInteger();
  Mockito.doAnswer(new Answer<Void>() {
    @Override
    public Void answer(InvocationOnMock invocation) throws Throwable {
      if (count.getAndIncrement() == 0) {
        throw new IOException("faked initBlockPool exception");
      }
      // The initBlockPool is called again. Now mock init is done.
      Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
      return null;
    }
  }).when(mockDn).initBlockPool(Mockito.any(BPOfferService.class));
  BPOfferService bpos = setupBPOSForNNs(mockDn, mockNN1, mockNN2);
  List<BPServiceActor> actors = bpos.getBPServiceActors();
  assertEquals(2, actors.size());
  bpos.start();
  try {
    waitForInitialization(bpos);
    // even if one of the actor initialization fails, the other one will be
    // finish block report.
    waitForBlockReport(mockNN1, mockNN2);
  } finally {
    bpos.stop();
  }
}
项目:hadoop-2.6.0-cdh5.4.3    文件:TestBPOfferService.java   
/**
 * Test datanode block pool initialization error handling.
 * Failure in initializing a block pool should not cause NPE.
 */
@Test
public void testBPInitErrorHandling() throws Exception {
  final DataNode mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(
    new File(TEST_BUILD_DATA, "testBPInitErrorHandling"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn")).
    when(mockDn).getMetrics();
  final AtomicInteger count = new AtomicInteger();
  Mockito.doAnswer(new Answer<Void>() {
    @Override
    public Void answer(InvocationOnMock invocation) throws Throwable {
      if (count.getAndIncrement() == 0) {
        throw new IOException("faked initBlockPool exception");
      }
      // The initBlockPool is called again. Now mock init is done.
      Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
      return null;
    }
  }).when(mockDn).initBlockPool(Mockito.any(BPOfferService.class));
  BPOfferService bpos = setupBPOSForNNs(mockDn, mockNN1, mockNN2);
  List<BPServiceActor> actors = bpos.getBPServiceActors();
  assertEquals(2, actors.size());
  bpos.start();
  try {
    waitForInitialization(bpos);
    // even if one of the actor initialization fails, the other one will be
    // finish block report.
    waitForBlockReport(mockNN1, mockNN2);
  } finally {
    bpos.stop();
  }
}
项目:hadoop-EAR    文件:TestStuckDataNode.java   
/** This creates a slow writer and check to see
  * if pipeline heartbeats work fine
  */
 public void testStuckDataNode() throws Exception {
   final int DATANODE_NUM = 3;
   Configuration conf = new Configuration();
   final int timeout = 8000;
   conf.setInt("dfs.socket.timeout",timeout);

   final Path p = new Path("/pipelineHeartbeat/foo");
   System.out.println("p=" + p);

   MiniDFSCluster cluster = new MiniDFSCluster(conf, DATANODE_NUM, true, null);
   DistributedFileSystem fs = (DistributedFileSystem)cluster.getFileSystem();

DataNodeMetrics metrics = cluster.getDataNodes().get(0).myMetrics;
MetricsTimeVaryingLong spyBytesWritten = spy(metrics.bytesWritten);
DelayAnswer delayAnswer = new DelayAnswer(); 
doAnswer(delayAnswer).when(spyBytesWritten).inc(anyInt());
metrics.bytesWritten = spyBytesWritten;

try {
    // create a new file.
    FSDataOutputStream stm = fs.create(p);
    stm.write(1);
    stm.sync();
    stm.write(2);
    stm.close();

    // verify that entire file is good
    FSDataInputStream in = fs.open(p);
    assertEquals(1, in.read());
    assertEquals(2, in.read());
    in.close();
   } finally {
     fs.close();
     cluster.shutdown();
   }
 }
项目:hadoop-plus    文件:DataNode.java   
/**
 * This method starts the data node with the specified conf.
 * 
 * @param conf - the configuration
 *  if conf's CONFIG_PROPERTY_SIMULATED property is set
 *  then a simulated storage based data node is created.
 * 
 * @param dataDirs - only for a non-simulated storage data node
 * @throws IOException
 */
void startDataNode(Configuration conf, 
                   AbstractList<File> dataDirs,
                  // DatanodeProtocol namenode,
                   SecureResources resources
                   ) throws IOException {
  if(UserGroupInformation.isSecurityEnabled() && resources == null) {
    if (!conf.getBoolean("ignore.secure.ports.for.testing", false)) {
      throw new RuntimeException("Cannot start secure cluster without "
          + "privileged resources.");
    }
  }

  // settings global for all BPs in the Data Node
  this.secureResources = resources;
  this.dataDirs = dataDirs;
  this.conf = conf;
  this.dnConf = new DNConf(conf);

  storage = new DataStorage();

  // global DN settings
  registerMXBean();
  initDataXceiver(conf);
  startInfoServer(conf);

  // BlockPoolTokenSecretManager is required to create ipc server.
  this.blockPoolTokenSecretManager = new BlockPoolTokenSecretManager();
  initIpcServer(conf);

  metrics = DataNodeMetrics.create(conf, getDisplayName());

  blockPoolManager = new BlockPoolManager(this);
  blockPoolManager.refreshNamenodes(conf);

  // Create the ReadaheadPool from the DataNode context so we can
  // exit without having to explicitly shutdown its thread pool.
  readaheadPool = ReadaheadPool.getInstance();
}
项目:PDHC    文件:CheckerNode.java   
/**
 * This method starts the data node with the specified conf.
 * 
 * @param conf - the configuration
 *  if conf's CONFIG_PROPERTY_SIMULATED property is set
 *  then a simulated storage based data node is created.
 * 
 * @param dataDirs - only for a non-simulated storage data node
 * @throws IOException
 */
void startDataNode(Configuration conf, 
                   AbstractList<File> dataDirs,
                  // DatanodeProtocol namenode,
                   SecureResources resources
                   ) throws IOException {
  if(UserGroupInformation.isSecurityEnabled() && resources == null) {
    if (!conf.getBoolean("ignore.secure.ports.for.testing", false)) {
      throw new RuntimeException("Cannot start secure cluster without "
          + "privileged resources.");
    }
  }

  // settings global for all BPs in the Data Node
  this.secureResources = resources;
  this.dataDirs = dataDirs;
  this.conf = conf;
  this.dnConf = new DNConf(conf);

  storage = new DataStorage();

  // global DN settings
  registerMXBean();
  initDataXceiver(conf);
  startInfoServer(conf);

  // BlockPoolTokenSecretManager is required to create ipc server.
  this.blockPoolTokenSecretManager = new BlockPoolTokenSecretManager();
  initIpcServer(conf);

  metrics = DataNodeMetrics.create(conf, getDisplayName());

  blockPoolManager = new BlockPoolManager(this);
  blockPoolManager.refreshNamenodes(conf);

  // Create the ReadaheadPool from the DataNode context so we can
  // exit without having to explicitly shutdown its thread pool.
  readaheadPool = ReadaheadPool.getInstance();
}
项目:FlexMap    文件:TestBPOfferService.java   
/**
 * Test datanode block pool initialization error handling.
 * Failure in initializing a block pool should not cause NPE.
 */
@Test
public void testBPInitErrorHandling() throws Exception {
  final DataNode mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(
    new File(TEST_BUILD_DATA, "testBPInitErrorHandling"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn")).
    when(mockDn).getMetrics();
  final AtomicInteger count = new AtomicInteger();
  Mockito.doAnswer(new Answer<Void>() {
    @Override
    public Void answer(InvocationOnMock invocation) throws Throwable {
      if (count.getAndIncrement() == 0) {
        throw new IOException("faked initBlockPool exception");
      }
      // The initBlockPool is called again. Now mock init is done.
      Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
      return null;
    }
  }).when(mockDn).initBlockPool(Mockito.any(BPOfferService.class));
  BPOfferService bpos = setupBPOSForNNs(mockDn, mockNN1, mockNN2);
  List<BPServiceActor> actors = bpos.getBPServiceActors();
  assertEquals(2, actors.size());
  bpos.start();
  try {
    waitForInitialization(bpos);
    // even if one of the actor initialization fails, the other one will be
    // finish block report.
    waitForBlockReport(mockNN1, mockNN2);
  } finally {
    bpos.stop();
  }
}
项目:hops    文件:DataNode.java   
/**
 * This method starts the data node with the specified conf.
 *
 * @param conf
 *     - the configuration
 *     if conf's CONFIG_PROPERTY_SIMULATED property is set
 *     then a simulated storage based data node is created.
 * @param dataDirs
 *     - only for a non-simulated storage data node
 * @throws IOException
 */
void startDataNode(Configuration conf, AbstractList<File> dataDirs,
    // DatanodeProtocol namenode,
    SecureResources resources) throws IOException {
  if (UserGroupInformation.isSecurityEnabled() && resources == null) {
    if (!conf.getBoolean("ignore.secure.ports.for.testing", false)) {
      throw new RuntimeException(
          "Cannot start secure cluster without " + "privileged resources.");
    }
  }

  // settings global for all BPs in the Data Node
  this.secureResources = resources;
  this.dataDirs = dataDirs;
  this.conf = conf;
  this.dnConf = new DNConf(conf);

  storage = new DataStorage();

  // global DN settings
  registerMXBean();
  initDataXceiver(conf);
  startInfoServer(conf);

  // BlockPoolTokenSecretManager is required to create ipc server.
  this.blockPoolTokenSecretManager = new BlockPoolTokenSecretManager();
  initIpcServer(conf);

  metrics = DataNodeMetrics.create(conf, getDisplayName());

  blockPoolManager = new BlockPoolManager(this);
  blockPoolManager.refreshNamenodes(conf);

  // Create the ReadaheadPool from the DataNode context so we can
  // exit without having to explicitly shutdown its thread pool.
  readaheadPool = ReadaheadPool.getInstance();
}
项目:hadoop-TCP    文件:DataNode.java   
/**
 * This method starts the data node with the specified conf.
 * 
 * @param conf - the configuration
 *  if conf's CONFIG_PROPERTY_SIMULATED property is set
 *  then a simulated storage based data node is created.
 * 
 * @param dataDirs - only for a non-simulated storage data node
 * @throws IOException
 */
void startDataNode(Configuration conf, 
                   AbstractList<File> dataDirs,
                  // DatanodeProtocol namenode,
                   SecureResources resources
                   ) throws IOException {
  if(UserGroupInformation.isSecurityEnabled() && resources == null) {
    if (!conf.getBoolean("ignore.secure.ports.for.testing", false)) {
      throw new RuntimeException("Cannot start secure cluster without "
          + "privileged resources.");
    }
  }

  // settings global for all BPs in the Data Node
  this.secureResources = resources;
  this.dataDirs = dataDirs;
  this.conf = conf;
  this.dnConf = new DNConf(conf);

  storage = new DataStorage();

  // global DN settings
  registerMXBean();
  initDataXceiver(conf);
  startInfoServer(conf);

  // BlockPoolTokenSecretManager is required to create ipc server.
  this.blockPoolTokenSecretManager = new BlockPoolTokenSecretManager();
  initIpcServer(conf);

  metrics = DataNodeMetrics.create(conf, getDisplayName());

  blockPoolManager = new BlockPoolManager(this);
  blockPoolManager.refreshNamenodes(conf);

  // Create the ReadaheadPool from the DataNode context so we can
  // exit without having to explicitly shutdown its thread pool.
  readaheadPool = ReadaheadPool.getInstance();
}
项目:hardfs    文件:DataNode.java   
/**
 * This method starts the data node with the specified conf.
 * 
 * @param conf - the configuration
 *  if conf's CONFIG_PROPERTY_SIMULATED property is set
 *  then a simulated storage based data node is created.
 * 
 * @param dataDirs - only for a non-simulated storage data node
 * @throws IOException
 */
void startDataNode(Configuration conf, 
                   AbstractList<File> dataDirs,
                  // DatanodeProtocol namenode,
                   SecureResources resources
                   ) throws IOException {
  if(UserGroupInformation.isSecurityEnabled() && resources == null) {
    if (!conf.getBoolean("ignore.secure.ports.for.testing", false)) {
      throw new RuntimeException("Cannot start secure cluster without "
          + "privileged resources.");
    }
  }

  // settings global for all BPs in the Data Node
  this.secureResources = resources;
  this.dataDirs = dataDirs;
  this.conf = conf;
  this.dnConf = new DNConf(conf);

  storage = new DataStorage();

  // global DN settings
  registerMXBean();
  initDataXceiver(conf);
  startInfoServer(conf);

  // BlockPoolTokenSecretManager is required to create ipc server.
  this.blockPoolTokenSecretManager = new BlockPoolTokenSecretManager();
  initIpcServer(conf);

  metrics = DataNodeMetrics.create(conf, getDisplayName());

  blockPoolManager = new BlockPoolManager(this);
  blockPoolManager.refreshNamenodes(conf);

  // Create the ReadaheadPool from the DataNode context so we can
  // exit without having to explicitly shutdown its thread pool.
  readaheadPool = ReadaheadPool.getInstance();
}
项目:hadoop-on-lustre2    文件:TestBPOfferService.java   
/**
 * Test datanode block pool initialization error handling.
 * Failure in initializing a block pool should not cause NPE.
 */
@Test
public void testBPInitErrorHandling() throws Exception {
  final DataNode mockDn = Mockito.mock(DataNode.class);
  Mockito.doReturn(true).when(mockDn).shouldRun();
  Configuration conf = new Configuration();
  File dnDataDir = new File(
    new File(TEST_BUILD_DATA, "testBPInitErrorHandling"), "data");
  conf.set(DFS_DATANODE_DATA_DIR_KEY, dnDataDir.toURI().toString());
  Mockito.doReturn(conf).when(mockDn).getConf();
  Mockito.doReturn(new DNConf(conf)).when(mockDn).getDnConf();
  Mockito.doReturn(DataNodeMetrics.create(conf, "fake dn")).
    when(mockDn).getMetrics();
  final AtomicInteger count = new AtomicInteger();
  Mockito.doAnswer(new Answer<Void>() {
    @Override
    public Void answer(InvocationOnMock invocation) throws Throwable {
      if (count.getAndIncrement() == 0) {
        throw new IOException("faked initBlockPool exception");
      }
      // The initBlockPool is called again. Now mock init is done.
      Mockito.doReturn(mockFSDataset).when(mockDn).getFSDataset();
      return null;
    }
  }).when(mockDn).initBlockPool(Mockito.any(BPOfferService.class));
  BPOfferService bpos = setupBPOSForNNs(mockDn, mockNN1, mockNN2);
  bpos.start();
  try {
    waitForInitialization(bpos);
    List<BPServiceActor> actors = bpos.getBPServiceActors();
    assertEquals(1, actors.size());
    BPServiceActor actor = actors.get(0);
    waitForBlockReport(actor.getNameNodeProxy());
  } finally {
    bpos.stop();
  }
}
项目:cumulus    文件:DataNode.java   
/**
 * This method starts the data node with the specified conf.
 * 
 * @param conf - the configuration
 *  if conf's CONFIG_PROPERTY_SIMULATED property is set
 *  then a simulated storage based data node is created.
 * 
 * @param dataDirs - only for a non-simulated storage data node
 * @throws IOException
 */
void startDataNode(Configuration conf, 
                   AbstractList<File> dataDirs,
                   DatanodeProtocol namenode, SecureResources resources
                   ) throws IOException {
  if(UserGroupInformation.isSecurityEnabled() && resources == null)
    throw new RuntimeException("Cannot start secure cluster without " +
    "privileged resources.");

  this.secureResources = resources;
  this.namenode = namenode;
  storage = new DataStorage();

  initConfig(conf);
  registerMXBean();
  initDataXceiver(conf);
  initFsDataSet(conf, dataDirs);
  initBlockScanner(conf);
  startInfoServer(conf);

  myMetrics = new DataNodeMetrics(conf, dnRegistration.getName());
  // TODO check what code removed here

  initIpcServer(conf);
  startPlugins(conf);

  // BlockTokenSecretManager is created here, but it shouldn't be
  // used until it is initialized in register().
  this.blockTokenSecretManager = new BlockTokenSecretManager(false, 0, 0);
}
项目:RDFS    文件:TestStuckDataNode.java   
/** This creates a slow writer and check to see
  * if pipeline heartbeats work fine
  */
 public void testStuckDataNode() throws Exception {
   final int DATANODE_NUM = 3;
   Configuration conf = new Configuration();
   final int timeout = 8000;
   conf.setInt("dfs.socket.timeout",timeout);

   final Path p = new Path("/pipelineHeartbeat/foo");
   System.out.println("p=" + p);

   MiniDFSCluster cluster = new MiniDFSCluster(conf, DATANODE_NUM, true, null);
   DistributedFileSystem fs = (DistributedFileSystem)cluster.getFileSystem();

DataNodeMetrics metrics = cluster.getDataNodes().get(0).myMetrics;
MetricsTimeVaryingLong spyBytesWritten = spy(metrics.bytesWritten);
DelayAnswer delayAnswer = new DelayAnswer(); 
doAnswer(delayAnswer).when(spyBytesWritten).inc(anyInt());
metrics.bytesWritten = spyBytesWritten;

try {
    // create a new file.
    FSDataOutputStream stm = fs.create(p);
    stm.write(1);
    stm.sync();
    stm.write(2);
    stm.close();

    // verify that entire file is good
    FSDataInputStream in = fs.open(p);
    assertEquals(1, in.read());
    assertEquals(2, in.read());
    in.close();
   } finally {
     fs.close();
     cluster.shutdown();
   }
 }
项目:hadoop    文件:DataNode.java   
public DataNodeMetrics getMetrics() {
  return metrics;
}
项目:aliyun-oss-hadoop-fs    文件:DataNode.java   
public DataNodeMetrics getMetrics() {
  return metrics;
}
项目:big-c    文件:DataNode.java   
public DataNodeMetrics getMetrics() {
  return metrics;
}
项目:hadoop-2.6.0-cdh5.4.3    文件:DataNode.java   
public DataNodeMetrics getMetrics() {
  return metrics;
}
项目:hadoop-EAR    文件:DataNode.java   
DataNodeMetrics getMetrics() {
  return myMetrics;
}
项目:hadoop-plus    文件:DataNode.java   
DataNodeMetrics getMetrics() {
  return metrics;
}
项目:PDHC    文件:CheckerNode.java   
DataNodeMetrics getMetrics() {
  return metrics;
}
项目:FlexMap    文件:DataNode.java   
public DataNodeMetrics getMetrics() {
  return metrics;
}
项目:hops    文件:DataNode.java   
DataNodeMetrics getMetrics() {
  return metrics;
}
项目:hadoop-TCP    文件:DataNode.java   
DataNodeMetrics getMetrics() {
  return metrics;
}
项目:hardfs    文件:DataNode.java   
DataNodeMetrics getMetrics() {
  return metrics;
}