| Line |
Stmt Id |
Pos |
Tree |
Symbol |
Tests |
Code |
|
50
|
1790
|
1837
-
1841
|
Literal
|
<nosymbol>
|
|
null
|
|
50
|
1804
|
1804
-
1804
|
Literal
|
<nosymbol>
|
|
()
|
|
50
|
1805
|
1804
-
1804
|
Block
|
<nosymbol>
|
|
()
|
|
51
|
1791
|
1882
-
1928
|
Select
|
org.apache.hadoop.hbase.NamespaceDescriptor.DEFAULT_NAMESPACE_NAME_STR
|
|
org.apache.hadoop.hbase.NamespaceDescriptor.DEFAULT_NAMESPACE_NAME_STR
|
|
51
|
1792
|
1853
-
1928
|
Apply
|
java.lang.Object.!=
|
|
name.getNamespaceAsString().!=(org.apache.hadoop.hbase.NamespaceDescriptor.DEFAULT_NAMESPACE_NAME_STR)
|
|
52
|
1793
|
1969
-
2014
|
Select
|
org.apache.hadoop.hbase.NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR
|
|
org.apache.hadoop.hbase.NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR
|
|
52
|
1794
|
1940
-
2014
|
Apply
|
java.lang.Object.!=
|
|
name.getNamespaceAsString().!=(org.apache.hadoop.hbase.NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR)
|
|
52
|
1800
|
1808
-
2118
|
Apply
|
scala.Boolean.&&
|
|
name.getNamespaceAsString().!=(null).&&(name.getNamespaceAsString().!=(org.apache.hadoop.hbase.NamespaceDescriptor.DEFAULT_NAMESPACE_NAME_STR)).&&(name.getNamespaceAsString().!=(org.apache.hadoop.hbase.NamespaceDescriptor.SYSTEM_NAMESPACE_NAME_STR)).&&(scala.util.Try.apply[Option[org.apache.hadoop.hbase.NamespaceDescriptor]](scala.Option.apply[org.apache.hadoop.hbase.NamespaceDescriptor](admin.getNamespaceDescriptor(name.getNamespaceAsString()))).getOrElse[Option[org.apache.hadoop.hbase.NamespaceDescriptor]](scala.None).isEmpty)
|
|
53
|
1795
|
2066
-
2091
|
Apply
|
org.apache.hadoop.hbase.TableName.getNamespaceAsString
|
|
name.getNamespaceAsString()
|
|
53
|
1796
|
2037
-
2092
|
Apply
|
org.apache.hadoop.hbase.client.Admin.getNamespaceDescriptor
|
|
admin.getNamespaceDescriptor(name.getNamespaceAsString())
|
|
53
|
1797
|
2030
-
2093
|
Apply
|
scala.Option.apply
|
|
scala.Option.apply[org.apache.hadoop.hbase.NamespaceDescriptor](admin.getNamespaceDescriptor(name.getNamespaceAsString()))
|
|
53
|
1798
|
2105
-
2109
|
Select
|
scala.None
|
|
scala.None
|
|
53
|
1799
|
2026
-
2118
|
Select
|
scala.Option.isEmpty
|
|
scala.util.Try.apply[Option[org.apache.hadoop.hbase.NamespaceDescriptor]](scala.Option.apply[org.apache.hadoop.hbase.NamespaceDescriptor](admin.getNamespaceDescriptor(name.getNamespaceAsString()))).getOrElse[Option[org.apache.hadoop.hbase.NamespaceDescriptor]](scala.None).isEmpty
|
|
54
|
1801
|
2150
-
2211
|
Apply
|
org.apache.hadoop.hbase.NamespaceDescriptor.Builder.build
|
|
org.apache.hadoop.hbase.NamespaceDescriptor.create(name.getNamespaceAsString()).build()
|
|
54
|
1802
|
2128
-
2212
|
Apply
|
org.apache.hadoop.hbase.client.Admin.createNamespace
|
|
admin.createNamespace(org.apache.hadoop.hbase.NamespaceDescriptor.create(name.getNamespaceAsString()).build())
|
|
54
|
1803
|
2128
-
2212
|
Block
|
org.apache.hadoop.hbase.client.Admin.createNamespace
|
|
admin.createNamespace(org.apache.hadoop.hbase.NamespaceDescriptor.create(name.getNamespaceAsString()).build())
|
|
57
|
1806
|
2278
-
2296
|
Literal
|
<nosymbol>
|
|
classOf[org.apache.hadoop.hbase.TableName]
|
|
57
|
1807
|
2241
-
2336
|
TypeApply
|
scala.Any.asInstanceOf
|
|
HBaseVersions.this.hTableDescriptorClass.getConstructor(classOf[org.apache.hadoop.hbase.TableName]).newInstance(name).asInstanceOf[AnyRef]
|
|
59
|
1818
|
2342
-
2720
|
Apply
|
scala.collection.IterableLike.foreach
|
|
colFamilies.foreach[Unit](((k: Array[Byte]) => {
val column: Any = HBaseVersions.this.hColumnDescriptorClass.getConstructor(classOf[[B]).newInstance(k);
bloom.foreach[Unit](((x$1: org.apache.hadoop.hbase.regionserver.BloomType) => HBaseVersions.this._setBloomFilterType.apply(column, x$1)));
compression.foreach[Unit](((x$2: org.apache.hadoop.hbase.io.compress.Compression.Algorithm) => HBaseVersions.this._setCompressionType.apply(column, x$2)));
encoding.foreach[Unit](((x$3: org.apache.hadoop.hbase.io.encoding.DataBlockEncoding) => HBaseVersions.this._setDataBlockEncoding.apply(column, x$3)));
inMemory.foreach[Unit](((x$4: Boolean) => HBaseVersions.this._setInMemory.apply(column, x$4)));
HBaseVersions.this._addFamily.apply(descriptor, column)
}))
|
|
60
|
1808
|
2388
-
2462
|
Apply
|
java.lang.reflect.Constructor.newInstance
|
|
HBaseVersions.this.hColumnDescriptorClass.getConstructor(classOf[[B]).newInstance(k)
|
|
61
|
1809
|
2483
-
2513
|
Apply
|
scala.Function2.apply
|
|
HBaseVersions.this._setBloomFilterType.apply(column, x$1)
|
|
61
|
1810
|
2469
-
2514
|
Apply
|
scala.Option.foreach
|
|
bloom.foreach[Unit](((x$1: org.apache.hadoop.hbase.regionserver.BloomType) => HBaseVersions.this._setBloomFilterType.apply(column, x$1)))
|
|
62
|
1811
|
2541
-
2571
|
Apply
|
scala.Function2.apply
|
|
HBaseVersions.this._setCompressionType.apply(column, x$2)
|
|
62
|
1812
|
2521
-
2572
|
Apply
|
scala.Option.foreach
|
|
compression.foreach[Unit](((x$2: org.apache.hadoop.hbase.io.compress.Compression.Algorithm) => HBaseVersions.this._setCompressionType.apply(column, x$2)))
|
|
63
|
1813
|
2596
-
2628
|
Apply
|
scala.Function2.apply
|
|
HBaseVersions.this._setDataBlockEncoding.apply(column, x$3)
|
|
63
|
1814
|
2579
-
2629
|
Apply
|
scala.Option.foreach
|
|
encoding.foreach[Unit](((x$3: org.apache.hadoop.hbase.io.encoding.DataBlockEncoding) => HBaseVersions.this._setDataBlockEncoding.apply(column, x$3)))
|
|
64
|
1815
|
2653
-
2676
|
Apply
|
scala.Function2.apply
|
|
HBaseVersions.this._setInMemory.apply(column, x$4)
|
|
64
|
1816
|
2636
-
2677
|
Apply
|
scala.Option.foreach
|
|
inMemory.foreach[Unit](((x$4: Boolean) => HBaseVersions.this._setInMemory.apply(column, x$4)))
|
|
65
|
1817
|
2684
-
2714
|
Apply
|
scala.Function2.apply
|
|
HBaseVersions.this._addFamily.apply(descriptor, column)
|
|
68
|
1825
|
2726
-
2860
|
Apply
|
scala.Option.foreach
|
|
coprocessor.foreach[Unit](((x0$1: (String, Option[org.apache.hadoop.fs.Path])) => x0$1 match {
case (_1: String, _2: Option[org.apache.hadoop.fs.Path])(String, Option[org.apache.hadoop.fs.Path])((clas @ _), (path @ _)) => HBaseVersions.this._addCoprocessor.apply(descriptor, clas, path.orNull[org.apache.hadoop.fs.Path](scala.Predef.$conforms[Null]), 1073741823, null)
}))
|
|
69
|
1819
|
2814
-
2814
|
TypeApply
|
scala.Predef.$conforms
|
|
scala.Predef.$conforms[Null]
|
|
69
|
1820
|
2809
-
2820
|
ApplyToImplicitArgs
|
scala.Option.orNull
|
|
path.orNull[org.apache.hadoop.fs.Path](scala.Predef.$conforms[Null])
|
|
69
|
1821
|
2822
-
2847
|
Literal
|
<nosymbol>
|
|
1073741823
|
|
69
|
1822
|
2849
-
2853
|
Literal
|
<nosymbol>
|
|
null
|
|
69
|
1823
|
2775
-
2854
|
Apply
|
scala.Function5.apply
|
|
HBaseVersions.this._addCoprocessor.apply(descriptor, clas, path.orNull[org.apache.hadoop.fs.Path](scala.Predef.$conforms[Null]), 1073741823, null)
|
|
69
|
1824
|
2775
-
2854
|
Block
|
scala.Function5.apply
|
|
HBaseVersions.this._addCoprocessor.apply(descriptor, clas, path.orNull[org.apache.hadoop.fs.Path](scala.Predef.$conforms[Null]), 1073741823, null)
|
|
72
|
1826
|
2907
-
2921
|
Select
|
scala.collection.SeqLike.isEmpty
|
|
splits.isEmpty
|
|
72
|
1827
|
2925
-
2929
|
Literal
|
<nosymbol>
|
|
null
|
|
72
|
1828
|
2925
-
2929
|
Block
|
<nosymbol>
|
|
null
|
|
72
|
1829
|
2939
-
2953
|
ApplyToImplicitArgs
|
scala.collection.TraversableOnce.toArray
|
|
splits.toArray[Array[Byte]]((ClassTag.apply[Array[Byte]](scala.runtime.ScalaRunTime.arrayClass(classOf[scala.Byte])): scala.reflect.ClassTag[Array[Byte]]))
|
|
72
|
1830
|
2939
-
2953
|
Block
|
scala.collection.TraversableOnce.toArray
|
|
splits.toArray[Array[Byte]]((ClassTag.apply[Array[Byte]](scala.runtime.ScalaRunTime.arrayClass(classOf[scala.Byte])): scala.reflect.ClassTag[Array[Byte]]))
|
|
72
|
1831
|
2866
-
2956
|
Apply
|
scala.Function3.apply
|
|
HBaseVersions.this._createTableAsync.apply(admin, descriptor, if (splits.isEmpty)
null
else
splits.toArray[Array[Byte]]((ClassTag.apply[Array[Byte]](scala.runtime.ScalaRunTime.arrayClass(classOf[scala.Byte])): scala.reflect.ClassTag[Array[Byte]])))
|
|
81
|
1832
|
3155
-
3187
|
Apply
|
scala.Function2.apply
|
|
HBaseVersions.this._disableTableAsync.apply(admin, table)
|
|
88
|
1833
|
3364
-
3380
|
Apply
|
scala.Function1.apply
|
|
HBaseVersions.this._available.apply(conf)
|
|
173
|
1834
|
7619
-
7634
|
Apply
|
java.lang.Class.getMethods
|
|
clas.getMethods()
|
|
173
|
1835
|
7640
-
7657
|
Apply
|
java.lang.Object.==
|
|
x$5.getName().==(name)
|
|
173
|
1837
|
7619
-
7767
|
Apply
|
scala.Option.getOrElse
|
|
scala.Predef.refArrayOps[java.lang.reflect.Method](clas.getMethods()).find(((x$5: java.lang.reflect.Method) => x$5.getName().==(name))).getOrElse[java.lang.reflect.Method](throw new java.lang.NoSuchMethodException(scala.StringContext.apply("Couldn\'t find ", ".", " method").s(clas.getSimpleName(), name)))
|
|
174
|
1836
|
7677
-
7761
|
Throw
|
<nosymbol>
|
|
throw new java.lang.NoSuchMethodException(scala.StringContext.apply("Couldn\'t find ", ".", " method").s(clas.getSimpleName(), name))
|
|
176
|
1838
|
7793
-
7817
|
Apply
|
java.lang.reflect.Method.getParameterTypes
|
|
method.getParameterTypes()
|
|
177
|
1839
|
7851
-
7852
|
Literal
|
<nosymbol>
|
|
1
|
|
177
|
1840
|
7856
-
7884
|
Apply
|
java.lang.Object.==
|
|
scala.Predef.refArrayOps[Class[_]](parameterTypes).head.==(param)
|
|
177
|
1841
|
7826
-
7884
|
Apply
|
scala.Boolean.&&
|
|
parameterTypes.length.==(1).&&(scala.Predef.refArrayOps[Class[_]](parameterTypes).head.==(param))
|
|
178
|
1848
|
7894
-
8056
|
Function
|
org.locationtech.geomesa.hbase.utils.HBaseVersions.$anonfun
|
|
((obj: Any, p: Any) => try {
method.invoke(obj, p.asInstanceOf[AnyRef]);
()
} catch {
case (e @ (_: java.lang.reflect.InvocationTargetException)) => throw e.getCause()
})
|
|
179
|
1842
|
7941
-
7963
|
TypeApply
|
scala.Any.asInstanceOf
|
|
p.asInstanceOf[AnyRef]
|
|
179
|
1843
|
7922
-
7964
|
Apply
|
java.lang.reflect.Method.invoke
|
|
method.invoke(obj, p.asInstanceOf[AnyRef])
|
|
179
|
1844
|
7935
-
7935
|
Literal
|
<nosymbol>
|
|
()
|
|
179
|
1845
|
7922
-
7964
|
Block
|
<nosymbol>
|
|
{
method.invoke(obj, p.asInstanceOf[AnyRef]);
()
}
|
|
180
|
1846
|
8022
-
8038
|
Throw
|
<nosymbol>
|
|
throw e.getCause()
|
|
180
|
1847
|
8022
-
8038
|
Block
|
<nosymbol>
|
|
throw e.getCause()
|
|
184
|
1849
|
8076
-
8202
|
Throw
|
<nosymbol>
|
|
throw new java.lang.NoSuchMethodException(scala.StringContext.apply("Couldn\'t find ", ".", " method with correct parameters: ", "").s(clas.getSimpleName(), name, method))
|
|
184
|
1850
|
8076
-
8202
|
Block
|
<nosymbol>
|
|
throw new java.lang.NoSuchMethodException(scala.StringContext.apply("Couldn\'t find ", ".", " method with correct parameters: ", "").s(clas.getSimpleName(), name, method))
|