this.mapFileBytesRead = getCounterValueAsLong(this.mapCounters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.FILES_BYTES_READ); this.mapFileBytesWritten = getCounterValueAsLong(this.mapCounters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.FILES_BYTES_WRITTEN); this.reduceFileBytesRead = getCounterValueAsLong(this.reduceCounters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.FILES_BYTES_READ); this.hdfsBytesRead = getCounterValueAsLong(this.counters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.HDFS_BYTES_READ); this.hdfsBytesWritten = getCounterValueAsLong(this.counters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.HDFS_BYTES_WRITTEN); this.mapSlotMillis = getCounterValueAsLong(this.counters, Constants.JOB_COUNTER_HADOOP2, Constants.SLOTS_MILLIS_MAPS); this.reduceSlotMillis = getCounterValueAsLong(this.counters, Constants.JOB_COUNTER_HADOOP2, Constants.SLOTS_MILLIS_REDUCES); this.reduceShuffleBytes = getCounterValueAsLong(this.reduceCounters, Constants.TASK_COUNTER_HADOOP2, Constants.REDUCE_SHUFFLE_BYTES); } else { // presume it's hadoop1 this.mapFileBytesRead = getCounterValueAsLong(this.mapCounters, Constants.FILESYSTEM_COUNTERS, Constants.FILES_BYTES_READ); this.mapFileBytesWritten = getCounterValueAsLong(this.mapCounters, Constants.FILESYSTEM_COUNTERS,
this.mapFileBytesRead = getCounterValueAsLong(this.mapCounters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.FILES_BYTES_READ); this.mapFileBytesWritten = getCounterValueAsLong(this.mapCounters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.FILES_BYTES_WRITTEN); this.reduceFileBytesRead = getCounterValueAsLong(this.reduceCounters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.FILES_BYTES_READ); this.hdfsBytesRead = getCounterValueAsLong(this.counters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.HDFS_BYTES_READ); this.hdfsBytesWritten = getCounterValueAsLong(this.counters, Constants.FILESYSTEM_COUNTER_HADOOP2, Constants.HDFS_BYTES_WRITTEN); this.mapSlotMillis = getCounterValueAsLong(this.counters, Constants.JOB_COUNTER_HADOOP2, Constants.SLOTS_MILLIS_MAPS); this.reduceSlotMillis = getCounterValueAsLong(this.counters, Constants.JOB_COUNTER_HADOOP2, Constants.SLOTS_MILLIS_REDUCES); this.reduceShuffleBytes = getCounterValueAsLong(this.reduceCounters, Constants.TASK_COUNTER_HADOOP2, Constants.REDUCE_SHUFFLE_BYTES); } else { // presume it's hadoop1 this.mapFileBytesRead = getCounterValueAsLong(this.mapCounters, Constants.FILESYSTEM_COUNTERS, Constants.FILES_BYTES_READ); this.mapFileBytesWritten = getCounterValueAsLong(this.mapCounters, Constants.FILESYSTEM_COUNTERS,
/** * test the get counter value function */ @Test public void testGetCounterValueAsLong() { CounterMap cm = new CounterMap(); String cg = Constants.FILESYSTEM_COUNTERS; String cname = Constants.FILES_BYTES_READ; Long expValue = 1234L; Counter c1 = new Counter(cg, cname, expValue); cm.add(c1); JobDetails jd = new JobDetails(null); assertEquals(expValue, jd.getCounterValueAsLong(cm, cg, cname)); // test non existent counter value Long zeroValue = 0L; assertEquals(zeroValue, jd.getCounterValueAsLong(cm, Constants.JOB_COUNTER_HADOOP2, Constants.SLOTS_MILLIS_MAPS)); }