--------------------------------------------------------------------------- FatalError Traceback (most recent call last) ~/miniconda3/envs/hail/lib/python3.7/site-packages/IPython/core/formatters.py in __call__(self, obj) 700 type_pprinters=self.type_printers, 701 deferred_pprinters=self.deferred_printers) --> 702 printer.pretty(obj) 703 printer.flush() 704 return stream.getvalue() ~/miniconda3/envs/hail/lib/python3.7/site-packages/IPython/lib/pretty.py in pretty(self, obj) 392 if cls is not object \ 393 and callable(cls.__dict__.get('__repr__')): --> 394 return _repr_pprint(obj, self, cycle) 395 396 return _default_pprint(obj, self, cycle) ~/miniconda3/envs/hail/lib/python3.7/site-packages/IPython/lib/pretty.py in _repr_pprint(obj, p, cycle) 698 """A pprint that just redirects to the normal repr function.""" 699 # Find newlines and replace them with p.break_() --> 700 output = repr(obj) 701 lines = output.splitlines() 702 with p.group(): ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/table.py in __repr__(self) 1295 1296 def __repr__(self): -> 1297 return self.__str__() 1298 1299 def data(self): ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/table.py in __str__(self) 1292 1293 def __str__(self): -> 1294 return self._ascii_str() 1295 1296 def __repr__(self): ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/table.py in _ascii_str(self) 1318 return s 1319 -> 1320 rows, has_more, dtype = self.data() 1321 fields = list(dtype) 1322 trunc_fields = [trunc(f) for f in fields] ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/table.py in data(self) 1302 row_dtype = t.row.dtype 1303 t = t.select(**{k: hl._showstr(v) for (k, v) in t.row.items()}) -> 1304 rows, has_more = t._take_n(self.n) 1305 self._data = (rows, has_more, row_dtype) 1306 return self._data ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/table.py in _take_n(self, n) 1449 has_more = False 1450 else: -> 1451 rows = self.take(n + 1) 1452 has_more = len(rows) > n 1453 rows = rows[:n] in take(self, n, _localize) ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/typecheck/check.py in wrapper(__original_func, *args, **kwargs) 575 def wrapper(__original_func, *args, **kwargs): 576 args_, kwargs_ = check_all(__original_func, args, kwargs, checkers, is_method=is_method) --> 577 return __original_func(*args_, **kwargs_) 578 579 return wrapper ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/table.py in take(self, n, _localize) 2119 """ 2120 -> 2121 return self.head(n).collect(_localize) 2122 2123 @typecheck_method(n=int) in collect(self, _localize) ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/typecheck/check.py in wrapper(__original_func, *args, **kwargs) 575 def wrapper(__original_func, *args, **kwargs): 576 args_, kwargs_ = check_all(__original_func, args, kwargs, checkers, is_method=is_method) --> 577 return __original_func(*args_, **kwargs_) 578 579 return wrapper ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/table.py in collect(self, _localize) 1918 e = construct_expr(rows_ir, hl.tarray(t.row.dtype)) 1919 if _localize: -> 1920 return Env.backend().execute(e._ir) 1921 else: 1922 return e ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/backend/py4j_backend.py in execute(self, ir, timed) 96 raise HailUserError(message_and_trace) from None 97 ---> 98 raise e ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/backend/py4j_backend.py in execute(self, ir, timed) 72 # print(self._hail_package.expr.ir.Pretty.apply(jir, True, -1)) 73 try: ---> 74 result = json.loads(self._jhc.backend().executeJSON(jir)) 75 value = ir.typ._from_json(result['value']) 76 timings = result['timings'] ~/miniconda3/envs/hail/lib/python3.7/site-packages/py4j/java_gateway.py in __call__(self, *args) 1303 answer = self.gateway_client.send_command(command) 1304 return_value = get_return_value( -> 1305 answer, self.gateway_client, self.target_id, self.name) 1306 1307 for temp_arg in temp_args: ~/miniconda3/envs/hail/lib/python3.7/site-packages/hail/backend/py4j_backend.py in deco(*args, **kwargs) 30 raise FatalError('%s\n\nJava stack trace:\n%s\n' 31 'Hail version: %s\n' ---> 32 'Error summary: %s' % (deepest, full, hail.__version__, deepest), error_id) from None 33 except pyspark.sql.utils.CapturedException as e: 34 raise FatalError('%s\n\nJava stack trace:\n%s\n' FatalError: HailException: array index out of bounds: index=0, length=0 ---------- Python traceback: File "", line 1, in mt_split = hl.experimental.sparse_split_multi(mt, filter_changed_loci=True) File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 200, in sparse_split_multi entries: ds[entries].map(transform_entries) File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 193, in transform_entries return hl.bind(with_local_a_index, lai) File "", line 2, in bind File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/expr/functions.py", line 507, in bind lambda_result = to_expr(f(*args)) File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 180, in with_local_a_index hl.range(0, 3).map(lambda i: hl.min( File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 183, in .map(lambda idx: old_entry.LPL[idx]))))) File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 183, in .map(lambda idx: old_entry.LPL[idx]))))) Java stack trace: org.apache.spark.SparkException: Job aborted due to stage failure: Task 2 in stage 0.0 failed 1 times, most recent failure: Lost task 2.0 in stage 0.0 (TID 2) (172.22.86.55 executor driver): is.hail.utils.HailException: array index out of bounds: index=0, length=0 ---------- Python traceback: File "", line 1, in mt_split = hl.experimental.sparse_split_multi(mt, filter_changed_loci=True) File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 200, in sparse_split_multi entries: ds[entries].map(transform_entries) File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 193, in transform_entries return hl.bind(with_local_a_index, lai) File "", line 2, in bind File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/expr/functions.py", line 507, in bind lambda_result = to_expr(f(*args)) File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 180, in with_local_a_index hl.range(0, 3).map(lambda i: hl.min( File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 183, in .map(lambda idx: old_entry.LPL[idx]))))) File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 183, in .map(lambda idx: old_entry.LPL[idx]))))) at __C278Compiled.__m341arrayref_bounds_check(Emit.scala:932) at __C278Compiled.applyregion664_677(Emit.scala) at __C278Compiled.applyregion18_704(Emit.scala) at __C278Compiled.apply(Emit.scala) at is.hail.expr.ir.TableMapRows.$anonfun$execute$43(TableIR.scala:1918) at scala.runtime.java8.JFunction1$mcJJ$sp.apply(JFunction1$mcJJ$sp.java:23) at scala.collection.Iterator$$anon$10.next(Iterator.scala:459) at scala.collection.Iterator$$anon$11.next(Iterator.scala:494) at is.hail.rvd.RVD$$anon$1.next(RVD.scala:214) at is.hail.rvd.RVD$$anon$1.next(RVD.scala:208) at scala.collection.Iterator$$anon$11.next(Iterator.scala:494) at scala.collection.Iterator$$anon$10.next(Iterator.scala:459) at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:512) at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:511) at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:511) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:488) at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:511) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:488) at scala.collection.Iterator.foreach(Iterator.scala:941) at scala.collection.Iterator.foreach$(Iterator.scala:941) at scala.collection.AbstractIterator.foreach(Iterator.scala:1429) at is.hail.expr.ir.Interpret$.itF$1(Interpret.scala:886) at is.hail.expr.ir.Interpret$.$anonfun$run$67(Interpret.scala:905) at is.hail.expr.ir.Interpret$.$anonfun$run$67$adapted(Interpret.scala:905) at is.hail.rvd.RVD.$anonfun$combine$1(RVD.scala:692) at is.hail.rvd.RVD.$anonfun$combine$1$adapted(RVD.scala:692) at is.hail.sparkextras.ContextRDD.$anonfun$cmapPartitionsWithIndex$2(ContextRDD.scala:259) at is.hail.sparkextras.ContextRDD.$anonfun$cmapPartitionsWithIndex$3(ContextRDD.scala:259) at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:484) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:490) at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:488) at is.hail.utils.richUtils.RichContextRDD$$anon$1.hasNext(RichContextRDD.scala:71) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:488) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:132) at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:52) at org.apache.spark.scheduler.Task.run(Task.scala:131) at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Driver stacktrace: at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2258) at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2207) at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2206) at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2206) at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1079) at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1079) at scala.Option.foreach(Option.scala:407) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1079) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2445) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2387) at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2376) at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:868) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2196) at org.apache.spark.SparkContext.runJob(SparkContext.scala:2291) at is.hail.rvd.RVD.combine(RVD.scala:725) at is.hail.expr.ir.Interpret$.run(Interpret.scala:904) at is.hail.expr.ir.Interpret$.alreadyLowered(Interpret.scala:53) at is.hail.expr.ir.InterpretNonCompilable$.interpretAndCoerce$1(InterpretNonCompilable.scala:16) at is.hail.expr.ir.InterpretNonCompilable$.rewrite$1(InterpretNonCompilable.scala:53) at is.hail.expr.ir.InterpretNonCompilable$.rewrite$1(InterpretNonCompilable.scala:39) at is.hail.expr.ir.InterpretNonCompilable$.apply(InterpretNonCompilable.scala:58) at is.hail.expr.ir.lowering.InterpretNonCompilablePass$.transform(LoweringPass.scala:67) at is.hail.expr.ir.lowering.LoweringPass.$anonfun$apply$3(LoweringPass.scala:15) at is.hail.utils.ExecutionTimer.time(ExecutionTimer.scala:81) at is.hail.expr.ir.lowering.LoweringPass.$anonfun$apply$1(LoweringPass.scala:15) at is.hail.utils.ExecutionTimer.time(ExecutionTimer.scala:81) at is.hail.expr.ir.lowering.LoweringPass.apply(LoweringPass.scala:13) at is.hail.expr.ir.lowering.LoweringPass.apply$(LoweringPass.scala:12) at is.hail.expr.ir.lowering.InterpretNonCompilablePass$.apply(LoweringPass.scala:62) at is.hail.expr.ir.lowering.LoweringPipeline.$anonfun$apply$1(LoweringPipeline.scala:14) at is.hail.expr.ir.lowering.LoweringPipeline.$anonfun$apply$1$adapted(LoweringPipeline.scala:12) at scala.collection.IndexedSeqOptimized.foreach(IndexedSeqOptimized.scala:36) at scala.collection.IndexedSeqOptimized.foreach$(IndexedSeqOptimized.scala:33) at scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:38) at is.hail.expr.ir.lowering.LoweringPipeline.apply(LoweringPipeline.scala:12) at is.hail.expr.ir.CompileAndEvaluate$._apply(CompileAndEvaluate.scala:28) at is.hail.backend.spark.SparkBackend._execute(SparkBackend.scala:378) at is.hail.backend.spark.SparkBackend.$anonfun$execute$1(SparkBackend.scala:362) at is.hail.expr.ir.ExecuteContext$.$anonfun$scoped$3(ExecuteContext.scala:47) at is.hail.utils.package$.using(package.scala:627) at is.hail.expr.ir.ExecuteContext$.$anonfun$scoped$2(ExecuteContext.scala:47) at is.hail.utils.package$.using(package.scala:627) at is.hail.annotations.RegionPool$.scoped(RegionPool.scala:13) at is.hail.expr.ir.ExecuteContext$.scoped(ExecuteContext.scala:46) at is.hail.backend.spark.SparkBackend.withExecuteContext(SparkBackend.scala:272) at is.hail.backend.spark.SparkBackend.execute(SparkBackend.scala:359) at is.hail.backend.spark.SparkBackend.$anonfun$executeJSON$1(SparkBackend.scala:403) at is.hail.utils.ExecutionTimer$.time(ExecutionTimer.scala:52) at is.hail.backend.spark.SparkBackend.executeJSON(SparkBackend.scala:401) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) at py4j.Gateway.invoke(Gateway.java:282) at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) at py4j.commands.CallCommand.execute(CallCommand.java:79) at py4j.GatewayConnection.run(GatewayConnection.java:238) at java.lang.Thread.run(Thread.java:748) is.hail.utils.HailException: array index out of bounds: index=0, length=0 ---------- Python traceback: File "", line 1, in mt_split = hl.experimental.sparse_split_multi(mt, filter_changed_loci=True) File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 200, in sparse_split_multi entries: ds[entries].map(transform_entries) File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 193, in transform_entries return hl.bind(with_local_a_index, lai) File "", line 2, in bind File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/expr/functions.py", line 507, in bind lambda_result = to_expr(f(*args)) File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 180, in with_local_a_index hl.range(0, 3).map(lambda i: hl.min( File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 183, in .map(lambda idx: old_entry.LPL[idx]))))) File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 183, in .map(lambda idx: old_entry.LPL[idx]))))) at __C278Compiled.__m341arrayref_bounds_check(Emit.scala:932) at __C278Compiled.applyregion664_677(Emit.scala) at __C278Compiled.applyregion18_704(Emit.scala) at __C278Compiled.apply(Emit.scala) at is.hail.expr.ir.TableMapRows.$anonfun$execute$43(TableIR.scala:1918) at scala.runtime.java8.JFunction1$mcJJ$sp.apply(JFunction1$mcJJ$sp.java:23) at scala.collection.Iterator$$anon$10.next(Iterator.scala:459) at scala.collection.Iterator$$anon$11.next(Iterator.scala:494) at is.hail.rvd.RVD$$anon$1.next(RVD.scala:214) at is.hail.rvd.RVD$$anon$1.next(RVD.scala:208) at scala.collection.Iterator$$anon$11.next(Iterator.scala:494) at scala.collection.Iterator$$anon$10.next(Iterator.scala:459) at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:512) at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:511) at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:511) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:488) at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:511) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:488) at scala.collection.Iterator.foreach(Iterator.scala:941) at scala.collection.Iterator.foreach$(Iterator.scala:941) at scala.collection.AbstractIterator.foreach(Iterator.scala:1429) at is.hail.expr.ir.Interpret$.itF$1(Interpret.scala:886) at is.hail.expr.ir.Interpret$.$anonfun$run$67(Interpret.scala:905) at is.hail.expr.ir.Interpret$.$anonfun$run$67$adapted(Interpret.scala:905) at is.hail.rvd.RVD.$anonfun$combine$1(RVD.scala:692) at is.hail.rvd.RVD.$anonfun$combine$1$adapted(RVD.scala:692) at is.hail.sparkextras.ContextRDD.$anonfun$cmapPartitionsWithIndex$2(ContextRDD.scala:259) at is.hail.sparkextras.ContextRDD.$anonfun$cmapPartitionsWithIndex$3(ContextRDD.scala:259) at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:484) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:490) at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:488) at is.hail.utils.richUtils.RichContextRDD$$anon$1.hasNext(RichContextRDD.scala:71) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:488) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:132) at org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:99) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:52) at org.apache.spark.scheduler.Task.run(Task.scala:131) at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:497) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1439) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:500) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Hail version: 0.2.67-bafea6b18247 Error summary: HailException: array index out of bounds: index=0, length=0 ---------- Python traceback: File "", line 1, in mt_split = hl.experimental.sparse_split_multi(mt, filter_changed_loci=True) File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 200, in sparse_split_multi entries: ds[entries].map(transform_entries) File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 193, in transform_entries return hl.bind(with_local_a_index, lai) File "", line 2, in bind File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/expr/functions.py", line 507, in bind lambda_result = to_expr(f(*args)) File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 180, in with_local_a_index hl.range(0, 3).map(lambda i: hl.min( File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 183, in .map(lambda idx: old_entry.LPL[idx]))))) File "", line 2, in map File "/home/antonio/miniconda3/envs/hail/lib/python3.7/site-packages/hail/experimental/vcf_combiner/sparse_split_multi.py", line 183, in .map(lambda idx: old_entry.LPL[idx])))))