Skip to content

kups.core.data

Batched

Mixin that validates consistent leading batch dimension across pytree leaves.

Source code in src/kups/core/data/batched.py
class Batched:
    """Mixin that validates consistent leading batch dimension across pytree leaves."""

    def __post_init__(self):
        """Validate that all array leaves share the same leading dimension.

        Raises:
            ValueError: If arrays have inconsistent or missing leading dimensions.
        """
        try:
            leaves = jax.tree.leaves(self)
            shapes = tuple(x.shape for x in leaves)
        except AttributeError:
            return
        if not isinstance(leaves[0], jax.Array | np.ndarray):
            return
        if len(shapes) > 0:
            reference = shapes[0]
            if len(reference) == 0:
                raise ValueError(
                    f"Batched cannot be initialized with no leading dimension. Got shapes: {shapes}."
                )
            errors = []
            for i, shape in enumerate(shapes[1:]):
                if len(shape) == 0 or shape[0] != reference[0]:
                    errors.append((i + 1, shape))
            if len(errors) > 0:
                msg = "Inconsistent shapes in batched data: "
                msg += f"Expected all leaves to have the same first dimension size {reference[0]}.\n"
                msg += "The following leaves have different shapes: "
                for idx, shape in errors:
                    msg += f"Leaf {idx} has shape {shape}, expected {reference}. "
                raise ValueError(msg)

    def __len__(self) -> int:
        """Return the batch size (leading dimension size).

        Returns:
            The size of the batch dimension across all arrays
        """
        return jax.tree.leaves(self)[0].shape[0]

    @property
    def size(self) -> int:
        """The batch size (same as len()).

        Returns:
            The size of the batch dimension across all arrays
        """
        return len(self)

size property

The batch size (same as len()).

Returns:

Type Description
int

The size of the batch dimension across all arrays

__len__()

Return the batch size (leading dimension size).

Returns:

Type Description
int

The size of the batch dimension across all arrays

Source code in src/kups/core/data/batched.py
def __len__(self) -> int:
    """Return the batch size (leading dimension size).

    Returns:
        The size of the batch dimension across all arrays
    """
    return jax.tree.leaves(self)[0].shape[0]

__post_init__()

Validate that all array leaves share the same leading dimension.

Raises:

Type Description
ValueError

If arrays have inconsistent or missing leading dimensions.

Source code in src/kups/core/data/batched.py
def __post_init__(self):
    """Validate that all array leaves share the same leading dimension.

    Raises:
        ValueError: If arrays have inconsistent or missing leading dimensions.
    """
    try:
        leaves = jax.tree.leaves(self)
        shapes = tuple(x.shape for x in leaves)
    except AttributeError:
        return
    if not isinstance(leaves[0], jax.Array | np.ndarray):
        return
    if len(shapes) > 0:
        reference = shapes[0]
        if len(reference) == 0:
            raise ValueError(
                f"Batched cannot be initialized with no leading dimension. Got shapes: {shapes}."
            )
        errors = []
        for i, shape in enumerate(shapes[1:]):
            if len(shape) == 0 or shape[0] != reference[0]:
                errors.append((i + 1, shape))
        if len(errors) > 0:
            msg = "Inconsistent shapes in batched data: "
            msg += f"Expected all leaves to have the same first dimension size {reference[0]}.\n"
            msg += "The following leaves have different shapes: "
            for idx, shape in errors:
                msg += f"Leaf {idx} has shape {shape}, expected {reference}. "
            raise ValueError(msg)

Buffered

Bases: Table[TLabel, TData], Generic[TLabel, TData]

:class:Table with buffer management for row occupation.

A Buffered[TKey, TData] IS-A :class:Table where some rows may be unoccupied (soft-deleted). The view function extracts an :class:Index leaf from the data whose :attr:~Index.valid_mask serves as the occupation mask: rows with OOB sentinel indices are considered unoccupied.

On construction, all leaves except the viewed leaf are sanitized: plain array leaves are zeroed and other :class:Index leaves get an OOB sentinel for unoccupied rows.

Attributes:

Name Type Description
view Callable[[TData], Index]

Static callable that extracts the authoritative Index leaf from the data.

Source code in src/kups/core/data/buffered.py
@dataclass
class Buffered(Table[TLabel, TData], Generic[TLabel, TData]):
    """:class:`Table` with buffer management for row occupation.

    A ``Buffered[TKey, TData]`` IS-A :class:`Table` where some rows may be
    unoccupied (soft-deleted).  The ``view`` function extracts an
    :class:`Index` leaf from the data whose :attr:`~Index.valid_mask` serves
    as the occupation mask: rows with OOB sentinel indices are considered
    unoccupied.

    On construction, all leaves *except* the viewed leaf are sanitized:
    plain array leaves are zeroed and other :class:`Index` leaves get an
    OOB sentinel for unoccupied rows.

    Attributes:
        view: Static callable that extracts the authoritative Index leaf
            from the data.
    """

    view: Callable[[TData], Index] = field(static=True)

    @property
    def occupation(self) -> Array:
        """Boolean mask derived from the viewed Index leaf's valid_mask."""
        return self.view(self.data).valid_mask

    @property
    def num_occupied(self) -> Array:
        """Number of occupied slots."""
        return self.occupation.sum()

    @skip_post_init_if_disabled
    def __post_init__(self):
        super().__post_init__()
        mask = self.occupation
        try:
            viewed_leaf = self.view(self.data)
        except AttributeError as e:
            raise ValueError(
                "Could not get occupation index with view. "
                "Most likely the user Buffered an item without .system "
                "attribute and did not overwrite `view`."
            ) from e
        assert isinstance(viewed_leaf, Index), (
            f"View must point towards an Index. Got {viewed_leaf}."
        )

        def _sanitize(leaf):
            if isinstance(leaf, Index):
                if leaf is viewed_leaf:
                    return leaf  # source of truth — do not sanitize
                oob_sentinel = len(leaf.keys)
                data = jnp.where(mask, leaf.indices, oob_sentinel)
                return Index(leaf.keys, data, leaf.max_count, _cls=leaf._cls)
            expand_axes = tuple(i for i in range(leaf.ndim) if i != 0)
            return jnp.where(
                jnp.expand_dims(mask, axis=expand_axes),
                leaf,
                jnp.zeros_like(leaf),
            )

        sanitized = jax.tree.map(
            _sanitize, self.data, is_leaf=lambda x: isinstance(x, Index)
        )
        object.__setattr__(self, "data", sanitized)

    def select_free(self, n: int) -> Index[TLabel]:
        """Return an ``Index`` referencing ``n`` unoccupied slots.

        Args:
            n: Number of free slots to select.

        Returns:
            ``Index`` of shape ``(n,)`` into this buffer's labels.
            If fewer than ``n`` free slots exist, excess entries use
            the OOB sentinel (``len(index)``).
        """
        data = jnp.where(~self.occupation, size=n, fill_value=len(self.keys))[0]
        return Index(self.keys, data)

    @classmethod
    def arange[L: SupportsSorting, D](
        cls,
        data: D,
        *,
        num_occupied: int | None = None,
        label: Callable[[int], L] = int,
        view: Callable[[D], Index] = system_view,
    ) -> Buffered[L, D]:
        """Create a ``Buffered`` with integer labels ``(0, 1, ..., n-1)``.

        The ``view`` function extracts the authoritative Index leaf from
        ``data``. If ``num_occupied`` is less than ``n``, the viewed leaf
        is masked so that trailing entries have OOB sentinel values.

        Args:
            data: Pytree of arrays with a common leading dimension.
            num_occupied: Number of leading slots marked as occupied.
                Defaults to all slots occupied.
            label: Callable mapping ``int`` to label values.
            view: Callable extracting the authoritative Index leaf.

        Returns:
            ``Buffered`` with labels ``(0, 1, ..., n-1)``.
        """
        n = jax.tree.leaves(data)[0].shape[0]
        n_occ = n if num_occupied is None else num_occupied
        index = tuple(map(label, range(n)))
        if n_occ < n:
            mask = jnp.arange(n) < n_occ
            viewed = view(data)
            masked = viewed.apply_mask(mask)
            data = bind(data, view).set(masked)
        return Buffered(
            index, data, view, _cls=label if isinstance(label, type) else None
        )

    @classmethod
    def full[D](
        cls, table: Table[TLabel, D], *, view: Callable[[D], Index] = system_view
    ) -> Buffered[TLabel, D]:
        """Create a fully-occupied ``Buffered`` from a ``Table``.

        Args:
            table: Source table.
            view: Callable extracting the authoritative Index leaf.

        Returns:
            ``Buffered`` with all rows marked as occupied.
        """
        return Buffered(table.keys, table.data, view)

    @classmethod
    def pad[L: int, D](
        cls,
        table: Table[L, D],
        num_free: int,
        *,
        view: Callable[[D], Index] = system_view,
    ) -> Buffered[L, D]:
        """Convert a ``Table`` to a ``Buffered`` with extra free rows.

        All original entries are marked as occupied. New labels are
        consecutive integers starting after the last existing label.
        Zero-padded data has OOB indices in the viewed Index leaf.

        Args:
            table: Source table (fully occupied in the result).
            num_free: Number of unoccupied rows to append.
            view: Callable extracting the authoritative Index leaf.

        Returns:
            ``Buffered`` with ``len(table) + num_free`` rows.
        """
        n = len(table)
        L_t = table.cls
        new_idx = tuple(map(L_t, range(n, num_free + n)))
        new_data = jax.tree.map(lambda x: pad_axis(x, (0, num_free), 0), table.data)
        # Ensure the viewed leaf has OOB for padded entries.
        if num_free > 0:
            mask = jnp.arange(n + num_free) < n
            viewed = view(new_data)
            masked = viewed.apply_mask(mask)
            new_data = bind(new_data, view).set(masked)
        return Buffered(table.keys + new_idx, new_data, view)

    def update[D](
        self: Buffered[TLabel, D], index: Index[TLabel], data: D, **kwargs
    ) -> Buffered[TLabel, D]:
        """Update rows, returning ``Buffered``.

        The viewed Index leaf in ``data`` must carry correct validity
        (OOB sentinel for unoccupied rows).
        """
        return cast(Buffered[TLabel, D], super().update(index, data, **kwargs))  # type: ignore

    def update_if[D, L: SupportsSorting](
        self: Buffered[TLabel, D],
        accept: Table[L, Array],
        indices: Index[TLabel],
        new_data: D,
    ) -> Buffered[TLabel, D]:
        """Conditionally update rows, returning ``Buffered``."""
        return cast(Buffered[TLabel, D], super().update_if(accept, indices, new_data))  # type: ignore

num_occupied property

Number of occupied slots.

occupation property

Boolean mask derived from the viewed Index leaf's valid_mask.

arange(data, *, num_occupied=None, label=int, view=system_view) classmethod

Create a Buffered with integer labels (0, 1, ..., n-1).

The view function extracts the authoritative Index leaf from data. If num_occupied is less than n, the viewed leaf is masked so that trailing entries have OOB sentinel values.

Parameters:

Name Type Description Default
data D

Pytree of arrays with a common leading dimension.

required
num_occupied int | None

Number of leading slots marked as occupied. Defaults to all slots occupied.

None
label Callable[[int], L]

Callable mapping int to label values.

int
view Callable[[D], Index]

Callable extracting the authoritative Index leaf.

system_view

Returns:

Type Description
Buffered[L, D]

Buffered with labels (0, 1, ..., n-1).

Source code in src/kups/core/data/buffered.py
@classmethod
def arange[L: SupportsSorting, D](
    cls,
    data: D,
    *,
    num_occupied: int | None = None,
    label: Callable[[int], L] = int,
    view: Callable[[D], Index] = system_view,
) -> Buffered[L, D]:
    """Create a ``Buffered`` with integer labels ``(0, 1, ..., n-1)``.

    The ``view`` function extracts the authoritative Index leaf from
    ``data``. If ``num_occupied`` is less than ``n``, the viewed leaf
    is masked so that trailing entries have OOB sentinel values.

    Args:
        data: Pytree of arrays with a common leading dimension.
        num_occupied: Number of leading slots marked as occupied.
            Defaults to all slots occupied.
        label: Callable mapping ``int`` to label values.
        view: Callable extracting the authoritative Index leaf.

    Returns:
        ``Buffered`` with labels ``(0, 1, ..., n-1)``.
    """
    n = jax.tree.leaves(data)[0].shape[0]
    n_occ = n if num_occupied is None else num_occupied
    index = tuple(map(label, range(n)))
    if n_occ < n:
        mask = jnp.arange(n) < n_occ
        viewed = view(data)
        masked = viewed.apply_mask(mask)
        data = bind(data, view).set(masked)
    return Buffered(
        index, data, view, _cls=label if isinstance(label, type) else None
    )

full(table, *, view=system_view) classmethod

Create a fully-occupied Buffered from a Table.

Parameters:

Name Type Description Default
table Table[TLabel, D]

Source table.

required
view Callable[[D], Index]

Callable extracting the authoritative Index leaf.

system_view

Returns:

Type Description
Buffered[TLabel, D]

Buffered with all rows marked as occupied.

Source code in src/kups/core/data/buffered.py
@classmethod
def full[D](
    cls, table: Table[TLabel, D], *, view: Callable[[D], Index] = system_view
) -> Buffered[TLabel, D]:
    """Create a fully-occupied ``Buffered`` from a ``Table``.

    Args:
        table: Source table.
        view: Callable extracting the authoritative Index leaf.

    Returns:
        ``Buffered`` with all rows marked as occupied.
    """
    return Buffered(table.keys, table.data, view)

pad(table, num_free, *, view=system_view) classmethod

Convert a Table to a Buffered with extra free rows.

All original entries are marked as occupied. New labels are consecutive integers starting after the last existing label. Zero-padded data has OOB indices in the viewed Index leaf.

Parameters:

Name Type Description Default
table Table[L, D]

Source table (fully occupied in the result).

required
num_free int

Number of unoccupied rows to append.

required
view Callable[[D], Index]

Callable extracting the authoritative Index leaf.

system_view

Returns:

Type Description
Buffered[L, D]

Buffered with len(table) + num_free rows.

Source code in src/kups/core/data/buffered.py
@classmethod
def pad[L: int, D](
    cls,
    table: Table[L, D],
    num_free: int,
    *,
    view: Callable[[D], Index] = system_view,
) -> Buffered[L, D]:
    """Convert a ``Table`` to a ``Buffered`` with extra free rows.

    All original entries are marked as occupied. New labels are
    consecutive integers starting after the last existing label.
    Zero-padded data has OOB indices in the viewed Index leaf.

    Args:
        table: Source table (fully occupied in the result).
        num_free: Number of unoccupied rows to append.
        view: Callable extracting the authoritative Index leaf.

    Returns:
        ``Buffered`` with ``len(table) + num_free`` rows.
    """
    n = len(table)
    L_t = table.cls
    new_idx = tuple(map(L_t, range(n, num_free + n)))
    new_data = jax.tree.map(lambda x: pad_axis(x, (0, num_free), 0), table.data)
    # Ensure the viewed leaf has OOB for padded entries.
    if num_free > 0:
        mask = jnp.arange(n + num_free) < n
        viewed = view(new_data)
        masked = viewed.apply_mask(mask)
        new_data = bind(new_data, view).set(masked)
    return Buffered(table.keys + new_idx, new_data, view)

select_free(n)

Return an Index referencing n unoccupied slots.

Parameters:

Name Type Description Default
n int

Number of free slots to select.

required

Returns:

Type Description
Index[TLabel]

Index of shape (n,) into this buffer's labels.

Index[TLabel]

If fewer than n free slots exist, excess entries use

Index[TLabel]

the OOB sentinel (len(index)).

Source code in src/kups/core/data/buffered.py
def select_free(self, n: int) -> Index[TLabel]:
    """Return an ``Index`` referencing ``n`` unoccupied slots.

    Args:
        n: Number of free slots to select.

    Returns:
        ``Index`` of shape ``(n,)`` into this buffer's labels.
        If fewer than ``n`` free slots exist, excess entries use
        the OOB sentinel (``len(index)``).
    """
    data = jnp.where(~self.occupation, size=n, fill_value=len(self.keys))[0]
    return Index(self.keys, data)

update(index, data, **kwargs)

Update rows, returning Buffered.

The viewed Index leaf in data must carry correct validity (OOB sentinel for unoccupied rows).

Source code in src/kups/core/data/buffered.py
def update[D](
    self: Buffered[TLabel, D], index: Index[TLabel], data: D, **kwargs
) -> Buffered[TLabel, D]:
    """Update rows, returning ``Buffered``.

    The viewed Index leaf in ``data`` must carry correct validity
    (OOB sentinel for unoccupied rows).
    """
    return cast(Buffered[TLabel, D], super().update(index, data, **kwargs))  # type: ignore

update_if(accept, indices, new_data)

Conditionally update rows, returning Buffered.

Source code in src/kups/core/data/buffered.py
def update_if[D, L: SupportsSorting](
    self: Buffered[TLabel, D],
    accept: Table[L, Array],
    indices: Index[TLabel],
    new_data: D,
) -> Buffered[TLabel, D]:
    """Conditionally update rows, returning ``Buffered``."""
    return cast(Buffered[TLabel, D], super().update_if(accept, indices, new_data))  # type: ignore

Index

JAX-compatible foreign-key column referencing a set of unique keys.

An Index[Key] stores a static key vocabulary (keys) and a JAX integer array of positions into that vocabulary (indices). This makes the column compatible with jax.jit while preserving categorical / relational semantics.

Attributes:

Name Type Description
keys tuple[Key, ...]

Unique sorted key vocabulary (stored as a static pytree field).

indices Array

Integer JAX array of positions into keys.

max_count int | None

Optional upper bound on occurrences per key.

Source code in src/kups/core/data/index.py
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
@dataclass
class Index[Key: SupportsSorting]:
    """JAX-compatible foreign-key column referencing a set of unique keys.

    An ``Index[Key]`` stores a static key vocabulary (``keys``) and a JAX
    integer array of positions into that vocabulary (``indices``).  This
    makes the column compatible with ``jax.jit`` while preserving
    categorical / relational semantics.

    Attributes:
        keys: Unique sorted key vocabulary (stored as a static pytree field).
        indices: Integer JAX array of positions into ``keys``.
        max_count: Optional upper bound on occurrences per key.
    """

    keys: tuple[Key, ...] = field(static=True)
    indices: Array
    max_count: int | None = field(static=True, default=None)
    _cls: type[Key] | None = field(static=True, default=None, kw_only=True)

    @property
    def dtype(self) -> jnp.dtype:
        return jnp.dtype(object)

    @property
    def shape(self) -> tuple[int, ...]:
        return self.indices.shape

    @property
    def size(self) -> int:
        return self.indices.size

    @property
    def ndim(self) -> int:
        return self.indices.ndim

    @property
    def T(self) -> "Index[Key]":
        return self.transpose()

    @property
    def cls(self) -> type[Key]:
        if self._cls is not None:
            return self._cls
        if len(self.keys) > 0:
            return type(self.keys[0])
        raise ValueError("Key class cannot be inferred.")

    @skip_post_init_if_disabled
    def __post_init__(self):
        if not isinstance(self.indices, Array):
            return
        assert jnp.issubdtype(self.indices.dtype, jnp.integer)
        np_data = np.empty(len(self.keys), dtype=object)
        np_data[:] = self.keys
        np_sorted, order = np.unique(np_data, return_inverse=True)
        if (np_sorted != np_data).any():
            raise ValueError("Keys must be unique and sorted.")
        object.__setattr__(self, "_cls", self.cls)

    @classmethod
    def new[T: SupportsSorting](
        cls, data: Sequence[T] | np.ndarray, *, max_count: int | None = None
    ) -> Index[T]:
        """Create an ``Index`` from a sequence of keys.

        Args:
            data: Input keys (any shape supported by ``np.asarray``).
            max_count: Optional upper bound on occurrences per key.
                If provided, validated against the actual data.

        Returns:
            A new ``Index`` with deduplicated sorted keys and
            integer-encoded data preserving the original shape.
        """
        np_data = np.asarray(data, dtype=object)
        unique_keys, values = np.unique(np_data, return_inverse=True)
        if max_count is not None and np_data.size > 0:
            _, counts = np.unique(values, return_counts=True)
            assert int(counts.max()) <= max_count, (
                f"max_count={max_count} exceeded: key appears {int(counts.max())} times"
            )
        key_tuple = tuple(unique_keys.tolist())
        return Index(
            key_tuple,
            jnp.asarray(values).reshape(np_data.shape),
            max_count,
            _cls=type(key_tuple[0]) if key_tuple else None,
        )

    @classmethod
    def arange[T: SupportsSorting](
        cls,
        n: int,
        label: Callable[[int], T] = int,
        max_count: int | None = None,
    ) -> Index[T]:
        """Create an ``Index`` with ``n`` unique keys, each appearing once.

        Equivalent to ``Index.integer(np.arange(n), n=n, label=label)``.

        Args:
            n: Number of keys (and elements).
            label: Callable mapping ``int`` to key values (default: ``int``).
            max_count: Optional upper bound on occurrences per key.
        """
        return cls.integer(np.arange(n), n=n, label=label, max_count=max_count)

    @classmethod
    def integer[T: SupportsSorting](
        cls,
        ids: Array | np.ndarray | Sequence[int],
        *,
        n: int | None = None,
        label: Callable[[int], T] = int,
        max_count: int | None = None,
    ) -> Index[T]:
        """Create an ``Index`` with keys ``label(0), ..., label(n-1)``.

        Args:
            ids: Integer array of key indices, shape arbitrary.
            n: Number of unique keys. If ``None``, inferred as
                ``int(ids.max()) + 1``.
            label: Callable mapping ``int`` to key values (default: ``int``).
            max_count: Optional upper bound on occurrences per key.
        """
        if not isinstance(ids, Array):
            ids = np.asarray(ids)
        if n is None:
            n = int(ids.max()) + 1
        return Index(
            tuple(label(i) for i in range(n)),
            jnp.asarray(ids),
            max_count,
            _cls=type(label(0)),
        )

    @classmethod
    def zeros[T: SupportsSorting](
        cls,
        shape: int | tuple[int, ...],
        *,
        label: Callable[[int], T] = int,
        max_count: int | None = None,
    ) -> Index[T]:
        """Create an ``Index`` filled with a single key (the zeroth).

        Args:
            shape: Shape of the resulting index array.
            label: Callable mapping ``int`` to key values (default: ``int``).
            max_count: Optional upper bound on occurrences per key.
        """
        return cls.integer(
            np.zeros(shape, dtype=int), n=1, label=label, max_count=max_count
        )

    @cached_property
    def value(self) -> np.ndarray:
        """Decoded numpy array of key values (same shape as ``indices``)."""
        return np.asarray(self.keys, dtype=object)[np.asarray(self.indices)]

    def __array__(self) -> np.ndarray:
        """Support ``np.asarray(index)`` by returning decoded keys."""
        return self.value

    def __iter__(self):
        """Iterate over elements, yielding scalar ``Index`` per entry."""
        return (
            Index(self.keys, i, max_count=self.max_count, _cls=self.cls)
            for i in self.indices
        )

    def __len__(self):
        """Number of elements along the first axis."""
        return len(self.indices)

    def __getitem__(self, item) -> Index[Key]:
        """Index into the underlying integer array, preserving keys."""
        return self._forward_to_data("__getitem__", item)

    @property
    def num_labels(self):
        """Number of unique keys in the vocabulary."""
        return len(self.keys)

    def transpose(self, *axes: int) -> Index[Key]:
        """Transpose the index array, preserving keys."""
        return self._forward_to_data("transpose", *axes)

    def ravel(self) -> Index[Key]:
        """Flatten to 1-D, preserving keys."""
        return self._forward_to_data("ravel")

    def reshape(self, *args, **kwargs) -> Index[Key]:
        """Reshape the index array, preserving keys."""
        return self._forward_to_data("reshape", *args, **kwargs)

    def _forward_to_data(self, name: str, *args, **kwargs) -> Index[Key]:
        return Index(
            self.keys,
            getattr(self.indices, name)(*args, **kwargs),
            self.max_count,
            _cls=self.cls,
        )

    @property
    def scatter_args(self) -> ScatterArgs:
        """Scatter args using ``len(keys)`` as OOB fill value."""
        return {"mode": "fill", "fill_value": len(self.keys)}

    @property
    def counts(self) -> Table[Key, Array]:
        """Number of occurrences of each key, as ``Table[Key, Array]``."""
        from kups.core.data.table import Table

        return Table(
            self.keys, jnp.bincount(self.indices.ravel(), length=len(self.keys))
        )

    @property
    def valid_mask(self) -> Array:
        """Boolean mask: ``True`` where the index is within bounds (not OOB)."""
        return self.indices < len(self.keys)

    def apply_mask(self, mask: Array) -> Index[Key]:
        """Set entries where ``mask`` is ``False`` to the OOB sentinel.

        Args:
            mask: Boolean array broadcastable to ``self.indices``.

        Returns:
            New ``Index`` with masked-out entries replaced by ``len(keys)``.
        """
        return Index(
            self.keys,
            jnp.where(mask, self.indices, len(self.keys)),
            max_count=self.max_count,
            _cls=self._cls,
        )

    def select_per_label(self, indices: Array) -> Array:
        """For each key, pick the k-th occurrence (relative to absolute index).

        Args:
            indices: Integer array of shape ``(len(keys),)`` with per-key
                relative indices. Values are taken modulo the key count.

        Returns:
            Integer array of shape ``(len(keys),)`` with absolute positions.
            Keys with zero occurrences return ``len(self)`` (OOB sentinel).
        """
        label_counts = self.counts.data
        indices = indices % jnp.maximum(label_counts, 1)
        sorted_indices = jnp.argsort(self.indices, stable=True)
        offsets = jnp.cumulative_sum(label_counts, include_initial=True)[:-1]
        result = sorted_indices.at[offsets + indices].get(
            mode="fill", fill_value=len(self)
        )
        return jnp.where(label_counts == 0, len(self), result)

    def where_rectangular(self, target: Index[Key], max_count: int) -> Array:
        """For each target key, find all positions padded to ``max_count``.

        Args:
            target: Index with keys to search for (must be a subset of ``self.keys``).
            max_count: Maximum number of positions per key.

        Returns:
            Integer array of shape ``(len(target), max_count)`` with positions
            for each target key. Excess entries filled with ``len(self)`` (OOB).
        """
        target_ids = target.indices_in(self.keys)

        @jax.vmap
        def _find(label: Array) -> Array:
            return jnp.where(
                self.indices == label, size=max_count, fill_value=len(self)
            )[0]

        return _find(target_ids)

    def where_flat(
        self, target: Index[Key], /, capacity: Capacity[int] | None = None
    ) -> Array:
        """Find all positions matching any target key, flat.

        Args:
            target: Index with keys to search for (must be a subset of ``self.keys``).
            capacity: Capacity controlling output buffer size.

        Returns:
            Integer array of matching positions. Excess entries are filled
            with ``len(self)`` (OOB sentinel).
        """
        target_ids = target.indices_in(self.keys)
        mask = isin(self.indices, target_ids, len(self.keys))
        required = mask.sum()
        if is_traced(required):
            assert capacity is not None, "Capacity must be set during tracing."
            size = capacity.generate_assertion(required).size
        else:
            size = None
        return jnp.where(mask, size=size, fill_value=len(self))[0]

    def indices_in(
        self, tokens: tuple[Key, ...], *, allow_missing: bool = False
    ) -> Array:
        """Map this array's elements to indices in a target key tuple.

        For each element in ``self``, returns the index of its key in
        ``tokens``. Useful for re-indexing into a different key ordering
        (e.g., mapping per-atom species to potential parameters).

        Args:
            tokens: Target key tuple whose elements must be a superset of
                ``self.keys`` (each key must appear exactly once).
            allow_missing: If ``True``, keys in ``self`` that are absent from
                ``tokens`` are mapped to the OOB sentinel instead of raising.

        Returns:
            A JAX integer array (same shape as ``self.indices``) containing the
            corresponding indices into ``tokens``.

        Raises:
            AssertionError: If any key in ``self`` is missing from ``tokens``.
        """
        if self.keys == tokens:
            return self.indices
        if self.keys == tokens[: len(self.keys)]:
            return jnp.where(self.indices == len(self.keys), len(tokens), self.indices)
        keys1 = np.asarray(self.keys, dtype=object)
        keys2 = np.asarray(tokens, dtype=object)
        mask = keys1[:, None] == keys2
        missing = keys1[~(mask.sum(-1) == 1).astype(bool)]
        if not allow_missing and len(missing) > 0:
            raise ValueError(f"Keys in self not found in target: {missing.tolist()}")
        if len(tokens) == 0:
            return jnp.full_like(self.indices, fill_value=0)
        idx_map = jnp.asarray(np.argmax(mask, axis=-1))
        return idx_map.at[self.indices].get(mode="fill", fill_value=len(tokens))

    def update_labels(
        self, labels: tuple[Key, ...], *, allow_missing: bool = False
    ) -> Index[Key]:
        """Re-index into a new key tuple, preserving logical mapping.

        Args:
            labels: New key tuple (must be a superset of ``self.keys``).
            allow_missing: If ``True``, keys in ``self`` that are absent from
                ``labels`` are mapped to the OOB sentinel instead of raising.

        Returns:
            New ``Index`` with ``labels`` and remapped indices.
        """
        return Index(
            labels,
            self.indices_in(labels, allow_missing=allow_missing),
            max_count=self.max_count,
            _cls=self.cls,
        )

    def subselect(
        self, needle: Index[Key], /, capacity: Capacity[int], is_sorted: bool = False
    ) -> IndexSubselectResult[Key]:
        """Find elements matching target keys, returning key-level indices.

        Args:
            needle: Index with target keys (must be a subset of ``self.keys``).
            capacity: Capacity controlling output buffer size.
            is_sorted: Whether ``self.indices`` is already sorted by key.

        Returns:
            :class:`IndexSubselectResult` with scatter/gather as Index[Key].
        """
        result = subselect(
            needle.indices_in(self.keys),
            self.indices,
            output_buffer_size=capacity,
            num_segments=len(self.keys),
            is_sorted=is_sorted,
        )
        return IndexSubselectResult(
            scatter=Index(
                needle.keys,
                needle.indices.at[result.scatter_idxs].get(**needle.scatter_args),
                _cls=needle.cls,
            ),
            gather=Index(
                self.keys,
                self.indices.at[result.gather_idxs].get(**self.scatter_args),
                _cls=self.cls,
            ),
        )

    def isin(self, other: Index[Key]) -> Array:
        """Test whether each element's key appears in ``other``'s keys.

        Args:
            other: Index whose keys define the membership set.

        Returns:
            Boolean JAX array (same shape as ``self.indices``).
        """
        all_keys = _merge_keys(self.keys, other.keys)
        lh_idx = self.indices_in(all_keys)
        rh_idx = other.indices_in(all_keys)
        max_item = len(all_keys)
        return isin(lh_idx, rh_idx, max_item)

    def __str__(self) -> str:
        keys_str = _format_keys(self.keys)
        return f"Index({self.cls.__name__}, keys={keys_str}, shape={self.shape})"

    def __repr__(self) -> str:
        keys_str = _format_keys(self.keys)
        try:
            strings = np.asarray(self.keys + ("OOB",))[np.asarray(self.indices)]
            arr_str = str(strings)
            return f"Index({arr_str}, cls={self.cls.__name__}, keys={keys_str}, shape={self.shape}, max_count={self.max_count})"
        except jax.errors.TracerArrayConversionError:
            return (
                f"Index(cls={self.cls.__name__}, keys={keys_str}, data={self.indices})"
            )

    def __tree_match__(self, *others: Index[Key]) -> tuple[Index[Key], ...]:
        all_indices: list[Index[Key]] = [self, *others]
        assert all(i.cls is self.cls for i in others), (
            f"cls mismatch: {[i.cls for i in all_indices]}"
        )
        new_keys = _merge_keys(*[t.keys for t in all_indices])
        max_count = max(
            (i.max_count for i in all_indices if i.max_count is not None), default=None
        )
        return tuple(
            Index(new_keys, i.indices_in(new_keys), max_count=max_count, _cls=i.cls)
            for i in all_indices
        )

    def to_cls[T: int, T2: SupportsSorting](
        self: Index[T], new: Callable[[int], T2] | type[T2]
    ) -> Index[T2]:
        """Convert keys to a different sentinel type.

        Args:
            new: Type or callable mapping each integer key to the new type.

        Returns:
            New ``Index`` with converted keys, same indices and max_count.
        """
        assert len(self.keys) > 0 or isinstance(new, type)
        new_keys = tuple(map(new, self.keys))
        cls = new if isinstance(new, type) else None
        return Index(new_keys, self.indices, max_count=self.max_count, _cls=cls)

    def sum_over(self, array: Array) -> Table[Key, Array]:
        """Sums ``array`` values grouped by this index via segment sum.

        Args:
            array: Array with leading dimension matching ``self.indices``.

        Returns:
            A ``Table`` mapping each key to its summed values.
        """
        from kups.core.data.table import Table

        return Table(
            self.keys,
            jax.ops.segment_sum(array, self.indices, self.num_labels, mode="drop"),
            _cls=self._cls,
        )

    @staticmethod
    def find[L: SupportsSorting](obj: PyTree, cls: type[L]) -> Index[L]:
        """Extracts the unique ``Index`` of a given type from a pytree.

        Args:
            obj: Pytree to search for ``Index`` leaves.
            cls: Key type to match against ``Index.cls``.

        Returns:
            The single ``Index`` leaf whose ``cls`` matches.

        Raises:
            AssertionError: If there is not exactly one matching ``Index``.
        """
        leaves = jax.tree.leaves(obj, is_leaf=lambda x: isinstance(x, Index))
        valid_leaves = [x for x in leaves if isinstance(x, Index) and x.cls is cls]
        assert len(valid_leaves) > 0, f"No Index[{cls.__name__}] found in pytree."
        assert len(valid_leaves) < 2, f"Multiple Index[{cls.__name__}] found in pytree."
        return valid_leaves[0]

    @overload
    @staticmethod
    def combine[L1: SupportsSorting](
        idx1: Index[L1],
        /,
    ) -> Index[tuple[L1]]: ...
    @overload
    @staticmethod
    def combine[L1: SupportsSorting, L2: SupportsSorting](
        idx1: Index[L1],
        idx2: Index[L2],
        /,
    ) -> Index[tuple[L1, L2]]: ...
    @overload
    @staticmethod
    def combine[L1: SupportsSorting, L2: SupportsSorting, L3: SupportsSorting](
        idx1: Index[L1],
        idx2: Index[L2],
        idx3: Index[L3],
        /,
    ) -> Index[tuple[L1, L2, L3]]: ...
    @overload
    @staticmethod
    def combine[
        L1: SupportsSorting,
        L2: SupportsSorting,
        L3: SupportsSorting,
        L4: SupportsSorting,
    ](
        idx1: Index[L1],
        idx2: Index[L2],
        idx3: Index[L3],
        idx4: Index[L4],
        /,
    ) -> Index[tuple[L1, L2, L3, L4]]: ...
    @staticmethod
    def combine(*indices: Index) -> Index:
        """Combines multiple indices into a single index of tuple keys.

        The key set is the full Cartesian product of all input key sets.

        Args:
            *indices: Indices to combine. Must all have the same shape.

        Returns:
            An ``Index`` whose keys are the Cartesian product of the
            input keys (as sorted tuples) and whose indices encode the
            per-element combination via mixed-radix encoding.

        Raises:
            AssertionError: If no indices are provided or shapes differ.

        Example::

            >>> idx1 = Index.new([ParticleId(0), ParticleId(1), ParticleId(0)])
            >>> idx2 = Index.new([SystemId(0), SystemId(0), SystemId(1)])
            >>> combined = Index.combine(idx1, idx2)
            >>> combined.keys  # 2 * 2 = 4 entries
            ((ParticleId(0), SystemId(0)), (ParticleId(0), SystemId(1)),
             (ParticleId(1), SystemId(0)), (ParticleId(1), SystemId(1)))
        """
        from itertools import product

        assert len(indices) > 0, "At least one index required."
        shape = indices[0].indices.shape
        assert all(idx.indices.shape == shape for idx in indices), (
            "Index shapes must match."
        )
        all_keys = tuple(sorted(product(*(idx.keys for idx in indices))))
        flat_key = jnp.zeros(indices[0].indices.size, dtype=jnp.int32)
        stride = 1
        for idx in reversed(indices):
            flat_key = flat_key + idx.indices.ravel() * stride
            stride *= idx.num_labels
        return Index(all_keys, flat_key.reshape(shape))

    @overload
    @staticmethod
    def concatenate[K: SupportsSorting](
        *indices: Index[K],
        shift_keys: Literal[False] = ...,
    ) -> Index[K]: ...
    @overload
    @staticmethod
    def concatenate[K: int](  # type: ignore[overload-overlap]
        *indices: Index[K],
        shift_keys: Literal[True],
    ) -> Index[K]: ...
    @staticmethod
    def concatenate(  # type: ignore[misc]
        *indices: Index,
        shift_keys: bool = False,
    ) -> Index:
        """Concatenate Index objects.

        Args:
            *indices: One or more Index objects to concatenate.
            shift_keys: If False (default), keys are merged (deduplicated,
                sorted) and indices remapped into the combined key space.
                If True, each input's keys are offset-shifted to be disjoint;
                requires integer sentinel keys (e.g. ``ParticleId``,
                ``SystemId``).

        Returns:
            A single concatenated Index.
        """
        assert len(indices) > 0, "At least 1 set of indices must be provided."
        assert all(i.cls == indices[0].cls for i in indices)

        if not shift_keys:
            new_keys = _merge_keys(*[i.keys for i in indices])
            parts = [i.indices_in(new_keys) if i.keys else i.indices for i in indices]
            new_indices = jnp.concatenate(parts) if parts else jnp.zeros(0, dtype=int)
            mc = 0
            for i in indices:
                if i.max_count is None:
                    mc = None
                    break
                mc += i.max_count
            return Index(new_keys, new_indices, mc, _cls=indices[0].cls)

        return _concatenate_shifted(*indices)  # type: ignore[arg-type]

    @overload
    @staticmethod
    def match[K: SupportsSorting](
        __i0: Index[K], __i1: Index[K], /
    ) -> tuple[Array, Array]: ...
    @overload
    @staticmethod
    def match[K: SupportsSorting](
        __i0: Index[K], __i1: Index[K], __i2: Index[K], /
    ) -> tuple[Array, Array, Array]: ...
    @overload
    @staticmethod
    def match[K: SupportsSorting](
        *indices: Index[K],
    ) -> tuple[Array, ...]: ...
    @staticmethod
    def match[K: SupportsSorting](*indices: Index[K]) -> tuple[Array, ...]:  # type: ignore[misc]
        """Remap multiple Index objects into a shared key space.

        Merges the key vocabularies of all inputs and returns the integer
        index arrays remapped into the combined key ordering. This is
        useful when two or more Index objects need element-wise comparison
        or alignment (e.g., matching species across systems).

        Args:
            *indices: One or more Index objects to align.

        Returns:
            A tuple of integer JAX arrays (one per input), each indexing
            into the merged key tuple.
        """
        all_keys = _merge_keys(*[i.keys for i in indices])
        return tuple(i.indices_in(all_keys) for i in indices)

counts property

Number of occurrences of each key, as Table[Key, Array].

num_labels property

Number of unique keys in the vocabulary.

scatter_args property

Scatter args using len(keys) as OOB fill value.

valid_mask property

Boolean mask: True where the index is within bounds (not OOB).

value cached property

Decoded numpy array of key values (same shape as indices).

__array__()

Support np.asarray(index) by returning decoded keys.

Source code in src/kups/core/data/index.py
def __array__(self) -> np.ndarray:
    """Support ``np.asarray(index)`` by returning decoded keys."""
    return self.value

__getitem__(item)

Index into the underlying integer array, preserving keys.

Source code in src/kups/core/data/index.py
def __getitem__(self, item) -> Index[Key]:
    """Index into the underlying integer array, preserving keys."""
    return self._forward_to_data("__getitem__", item)

__iter__()

Iterate over elements, yielding scalar Index per entry.

Source code in src/kups/core/data/index.py
def __iter__(self):
    """Iterate over elements, yielding scalar ``Index`` per entry."""
    return (
        Index(self.keys, i, max_count=self.max_count, _cls=self.cls)
        for i in self.indices
    )

__len__()

Number of elements along the first axis.

Source code in src/kups/core/data/index.py
def __len__(self):
    """Number of elements along the first axis."""
    return len(self.indices)

apply_mask(mask)

Set entries where mask is False to the OOB sentinel.

Parameters:

Name Type Description Default
mask Array

Boolean array broadcastable to self.indices.

required

Returns:

Type Description
Index[Key]

New Index with masked-out entries replaced by len(keys).

Source code in src/kups/core/data/index.py
def apply_mask(self, mask: Array) -> Index[Key]:
    """Set entries where ``mask`` is ``False`` to the OOB sentinel.

    Args:
        mask: Boolean array broadcastable to ``self.indices``.

    Returns:
        New ``Index`` with masked-out entries replaced by ``len(keys)``.
    """
    return Index(
        self.keys,
        jnp.where(mask, self.indices, len(self.keys)),
        max_count=self.max_count,
        _cls=self._cls,
    )

arange(n, label=int, max_count=None) classmethod

Create an Index with n unique keys, each appearing once.

Equivalent to Index.integer(np.arange(n), n=n, label=label).

Parameters:

Name Type Description Default
n int

Number of keys (and elements).

required
label Callable[[int], T]

Callable mapping int to key values (default: int).

int
max_count int | None

Optional upper bound on occurrences per key.

None
Source code in src/kups/core/data/index.py
@classmethod
def arange[T: SupportsSorting](
    cls,
    n: int,
    label: Callable[[int], T] = int,
    max_count: int | None = None,
) -> Index[T]:
    """Create an ``Index`` with ``n`` unique keys, each appearing once.

    Equivalent to ``Index.integer(np.arange(n), n=n, label=label)``.

    Args:
        n: Number of keys (and elements).
        label: Callable mapping ``int`` to key values (default: ``int``).
        max_count: Optional upper bound on occurrences per key.
    """
    return cls.integer(np.arange(n), n=n, label=label, max_count=max_count)

combine(*indices) staticmethod

combine(idx1: Index[L1]) -> Index[tuple[L1]]
combine(
    idx1: Index[L1], idx2: Index[L2]
) -> Index[tuple[L1, L2]]
combine(
    idx1: Index[L1], idx2: Index[L2], idx3: Index[L3]
) -> Index[tuple[L1, L2, L3]]
combine(
    idx1: Index[L1],
    idx2: Index[L2],
    idx3: Index[L3],
    idx4: Index[L4],
) -> Index[tuple[L1, L2, L3, L4]]

Combines multiple indices into a single index of tuple keys.

The key set is the full Cartesian product of all input key sets.

Parameters:

Name Type Description Default
*indices Index

Indices to combine. Must all have the same shape.

()

Returns:

Type Description
Index

An Index whose keys are the Cartesian product of the

Index

input keys (as sorted tuples) and whose indices encode the

Index

per-element combination via mixed-radix encoding.

Raises:

Type Description
AssertionError

If no indices are provided or shapes differ.

Example::

>>> idx1 = Index.new([ParticleId(0), ParticleId(1), ParticleId(0)])
>>> idx2 = Index.new([SystemId(0), SystemId(0), SystemId(1)])
>>> combined = Index.combine(idx1, idx2)
>>> combined.keys  # 2 * 2 = 4 entries
((ParticleId(0), SystemId(0)), (ParticleId(0), SystemId(1)),
 (ParticleId(1), SystemId(0)), (ParticleId(1), SystemId(1)))
Source code in src/kups/core/data/index.py
@staticmethod
def combine(*indices: Index) -> Index:
    """Combines multiple indices into a single index of tuple keys.

    The key set is the full Cartesian product of all input key sets.

    Args:
        *indices: Indices to combine. Must all have the same shape.

    Returns:
        An ``Index`` whose keys are the Cartesian product of the
        input keys (as sorted tuples) and whose indices encode the
        per-element combination via mixed-radix encoding.

    Raises:
        AssertionError: If no indices are provided or shapes differ.

    Example::

        >>> idx1 = Index.new([ParticleId(0), ParticleId(1), ParticleId(0)])
        >>> idx2 = Index.new([SystemId(0), SystemId(0), SystemId(1)])
        >>> combined = Index.combine(idx1, idx2)
        >>> combined.keys  # 2 * 2 = 4 entries
        ((ParticleId(0), SystemId(0)), (ParticleId(0), SystemId(1)),
         (ParticleId(1), SystemId(0)), (ParticleId(1), SystemId(1)))
    """
    from itertools import product

    assert len(indices) > 0, "At least one index required."
    shape = indices[0].indices.shape
    assert all(idx.indices.shape == shape for idx in indices), (
        "Index shapes must match."
    )
    all_keys = tuple(sorted(product(*(idx.keys for idx in indices))))
    flat_key = jnp.zeros(indices[0].indices.size, dtype=jnp.int32)
    stride = 1
    for idx in reversed(indices):
        flat_key = flat_key + idx.indices.ravel() * stride
        stride *= idx.num_labels
    return Index(all_keys, flat_key.reshape(shape))

concatenate(*indices, shift_keys=False) staticmethod

concatenate(
    *indices: Index[K], shift_keys: Literal[False] = ...
) -> Index[K]
concatenate(
    *indices: Index[K], shift_keys: Literal[True]
) -> Index[K]

Concatenate Index objects.

Parameters:

Name Type Description Default
*indices Index

One or more Index objects to concatenate.

()
shift_keys bool

If False (default), keys are merged (deduplicated, sorted) and indices remapped into the combined key space. If True, each input's keys are offset-shifted to be disjoint; requires integer sentinel keys (e.g. ParticleId, SystemId).

False

Returns:

Type Description
Index

A single concatenated Index.

Source code in src/kups/core/data/index.py
@staticmethod
def concatenate(  # type: ignore[misc]
    *indices: Index,
    shift_keys: bool = False,
) -> Index:
    """Concatenate Index objects.

    Args:
        *indices: One or more Index objects to concatenate.
        shift_keys: If False (default), keys are merged (deduplicated,
            sorted) and indices remapped into the combined key space.
            If True, each input's keys are offset-shifted to be disjoint;
            requires integer sentinel keys (e.g. ``ParticleId``,
            ``SystemId``).

    Returns:
        A single concatenated Index.
    """
    assert len(indices) > 0, "At least 1 set of indices must be provided."
    assert all(i.cls == indices[0].cls for i in indices)

    if not shift_keys:
        new_keys = _merge_keys(*[i.keys for i in indices])
        parts = [i.indices_in(new_keys) if i.keys else i.indices for i in indices]
        new_indices = jnp.concatenate(parts) if parts else jnp.zeros(0, dtype=int)
        mc = 0
        for i in indices:
            if i.max_count is None:
                mc = None
                break
            mc += i.max_count
        return Index(new_keys, new_indices, mc, _cls=indices[0].cls)

    return _concatenate_shifted(*indices)  # type: ignore[arg-type]

find(obj, cls) staticmethod

Extracts the unique Index of a given type from a pytree.

Parameters:

Name Type Description Default
obj PyTree

Pytree to search for Index leaves.

required
cls type[L]

Key type to match against Index.cls.

required

Returns:

Type Description
Index[L]

The single Index leaf whose cls matches.

Raises:

Type Description
AssertionError

If there is not exactly one matching Index.

Source code in src/kups/core/data/index.py
@staticmethod
def find[L: SupportsSorting](obj: PyTree, cls: type[L]) -> Index[L]:
    """Extracts the unique ``Index`` of a given type from a pytree.

    Args:
        obj: Pytree to search for ``Index`` leaves.
        cls: Key type to match against ``Index.cls``.

    Returns:
        The single ``Index`` leaf whose ``cls`` matches.

    Raises:
        AssertionError: If there is not exactly one matching ``Index``.
    """
    leaves = jax.tree.leaves(obj, is_leaf=lambda x: isinstance(x, Index))
    valid_leaves = [x for x in leaves if isinstance(x, Index) and x.cls is cls]
    assert len(valid_leaves) > 0, f"No Index[{cls.__name__}] found in pytree."
    assert len(valid_leaves) < 2, f"Multiple Index[{cls.__name__}] found in pytree."
    return valid_leaves[0]

indices_in(tokens, *, allow_missing=False)

Map this array's elements to indices in a target key tuple.

For each element in self, returns the index of its key in tokens. Useful for re-indexing into a different key ordering (e.g., mapping per-atom species to potential parameters).

Parameters:

Name Type Description Default
tokens tuple[Key, ...]

Target key tuple whose elements must be a superset of self.keys (each key must appear exactly once).

required
allow_missing bool

If True, keys in self that are absent from tokens are mapped to the OOB sentinel instead of raising.

False

Returns:

Type Description
Array

A JAX integer array (same shape as self.indices) containing the

Array

corresponding indices into tokens.

Raises:

Type Description
AssertionError

If any key in self is missing from tokens.

Source code in src/kups/core/data/index.py
def indices_in(
    self, tokens: tuple[Key, ...], *, allow_missing: bool = False
) -> Array:
    """Map this array's elements to indices in a target key tuple.

    For each element in ``self``, returns the index of its key in
    ``tokens``. Useful for re-indexing into a different key ordering
    (e.g., mapping per-atom species to potential parameters).

    Args:
        tokens: Target key tuple whose elements must be a superset of
            ``self.keys`` (each key must appear exactly once).
        allow_missing: If ``True``, keys in ``self`` that are absent from
            ``tokens`` are mapped to the OOB sentinel instead of raising.

    Returns:
        A JAX integer array (same shape as ``self.indices``) containing the
        corresponding indices into ``tokens``.

    Raises:
        AssertionError: If any key in ``self`` is missing from ``tokens``.
    """
    if self.keys == tokens:
        return self.indices
    if self.keys == tokens[: len(self.keys)]:
        return jnp.where(self.indices == len(self.keys), len(tokens), self.indices)
    keys1 = np.asarray(self.keys, dtype=object)
    keys2 = np.asarray(tokens, dtype=object)
    mask = keys1[:, None] == keys2
    missing = keys1[~(mask.sum(-1) == 1).astype(bool)]
    if not allow_missing and len(missing) > 0:
        raise ValueError(f"Keys in self not found in target: {missing.tolist()}")
    if len(tokens) == 0:
        return jnp.full_like(self.indices, fill_value=0)
    idx_map = jnp.asarray(np.argmax(mask, axis=-1))
    return idx_map.at[self.indices].get(mode="fill", fill_value=len(tokens))

integer(ids, *, n=None, label=int, max_count=None) classmethod

Create an Index with keys label(0), ..., label(n-1).

Parameters:

Name Type Description Default
ids Array | ndarray | Sequence[int]

Integer array of key indices, shape arbitrary.

required
n int | None

Number of unique keys. If None, inferred as int(ids.max()) + 1.

None
label Callable[[int], T]

Callable mapping int to key values (default: int).

int
max_count int | None

Optional upper bound on occurrences per key.

None
Source code in src/kups/core/data/index.py
@classmethod
def integer[T: SupportsSorting](
    cls,
    ids: Array | np.ndarray | Sequence[int],
    *,
    n: int | None = None,
    label: Callable[[int], T] = int,
    max_count: int | None = None,
) -> Index[T]:
    """Create an ``Index`` with keys ``label(0), ..., label(n-1)``.

    Args:
        ids: Integer array of key indices, shape arbitrary.
        n: Number of unique keys. If ``None``, inferred as
            ``int(ids.max()) + 1``.
        label: Callable mapping ``int`` to key values (default: ``int``).
        max_count: Optional upper bound on occurrences per key.
    """
    if not isinstance(ids, Array):
        ids = np.asarray(ids)
    if n is None:
        n = int(ids.max()) + 1
    return Index(
        tuple(label(i) for i in range(n)),
        jnp.asarray(ids),
        max_count,
        _cls=type(label(0)),
    )

isin(other)

Test whether each element's key appears in other's keys.

Parameters:

Name Type Description Default
other Index[Key]

Index whose keys define the membership set.

required

Returns:

Type Description
Array

Boolean JAX array (same shape as self.indices).

Source code in src/kups/core/data/index.py
def isin(self, other: Index[Key]) -> Array:
    """Test whether each element's key appears in ``other``'s keys.

    Args:
        other: Index whose keys define the membership set.

    Returns:
        Boolean JAX array (same shape as ``self.indices``).
    """
    all_keys = _merge_keys(self.keys, other.keys)
    lh_idx = self.indices_in(all_keys)
    rh_idx = other.indices_in(all_keys)
    max_item = len(all_keys)
    return isin(lh_idx, rh_idx, max_item)

match(*indices) staticmethod

match(
    __i0: Index[K], __i1: Index[K]
) -> tuple[Array, Array]
match(
    __i0: Index[K], __i1: Index[K], __i2: Index[K]
) -> tuple[Array, Array, Array]
match(*indices: Index[K]) -> tuple[Array, ...]

Remap multiple Index objects into a shared key space.

Merges the key vocabularies of all inputs and returns the integer index arrays remapped into the combined key ordering. This is useful when two or more Index objects need element-wise comparison or alignment (e.g., matching species across systems).

Parameters:

Name Type Description Default
*indices Index[K]

One or more Index objects to align.

()

Returns:

Type Description
Array

A tuple of integer JAX arrays (one per input), each indexing

...

into the merged key tuple.

Source code in src/kups/core/data/index.py
@staticmethod
def match[K: SupportsSorting](*indices: Index[K]) -> tuple[Array, ...]:  # type: ignore[misc]
    """Remap multiple Index objects into a shared key space.

    Merges the key vocabularies of all inputs and returns the integer
    index arrays remapped into the combined key ordering. This is
    useful when two or more Index objects need element-wise comparison
    or alignment (e.g., matching species across systems).

    Args:
        *indices: One or more Index objects to align.

    Returns:
        A tuple of integer JAX arrays (one per input), each indexing
        into the merged key tuple.
    """
    all_keys = _merge_keys(*[i.keys for i in indices])
    return tuple(i.indices_in(all_keys) for i in indices)

new(data, *, max_count=None) classmethod

Create an Index from a sequence of keys.

Parameters:

Name Type Description Default
data Sequence[T] | ndarray

Input keys (any shape supported by np.asarray).

required
max_count int | None

Optional upper bound on occurrences per key. If provided, validated against the actual data.

None

Returns:

Type Description
Index[T]

A new Index with deduplicated sorted keys and

Index[T]

integer-encoded data preserving the original shape.

Source code in src/kups/core/data/index.py
@classmethod
def new[T: SupportsSorting](
    cls, data: Sequence[T] | np.ndarray, *, max_count: int | None = None
) -> Index[T]:
    """Create an ``Index`` from a sequence of keys.

    Args:
        data: Input keys (any shape supported by ``np.asarray``).
        max_count: Optional upper bound on occurrences per key.
            If provided, validated against the actual data.

    Returns:
        A new ``Index`` with deduplicated sorted keys and
        integer-encoded data preserving the original shape.
    """
    np_data = np.asarray(data, dtype=object)
    unique_keys, values = np.unique(np_data, return_inverse=True)
    if max_count is not None and np_data.size > 0:
        _, counts = np.unique(values, return_counts=True)
        assert int(counts.max()) <= max_count, (
            f"max_count={max_count} exceeded: key appears {int(counts.max())} times"
        )
    key_tuple = tuple(unique_keys.tolist())
    return Index(
        key_tuple,
        jnp.asarray(values).reshape(np_data.shape),
        max_count,
        _cls=type(key_tuple[0]) if key_tuple else None,
    )

ravel()

Flatten to 1-D, preserving keys.

Source code in src/kups/core/data/index.py
def ravel(self) -> Index[Key]:
    """Flatten to 1-D, preserving keys."""
    return self._forward_to_data("ravel")

reshape(*args, **kwargs)

Reshape the index array, preserving keys.

Source code in src/kups/core/data/index.py
def reshape(self, *args, **kwargs) -> Index[Key]:
    """Reshape the index array, preserving keys."""
    return self._forward_to_data("reshape", *args, **kwargs)

select_per_label(indices)

For each key, pick the k-th occurrence (relative to absolute index).

Parameters:

Name Type Description Default
indices Array

Integer array of shape (len(keys),) with per-key relative indices. Values are taken modulo the key count.

required

Returns:

Type Description
Array

Integer array of shape (len(keys),) with absolute positions.

Array

Keys with zero occurrences return len(self) (OOB sentinel).

Source code in src/kups/core/data/index.py
def select_per_label(self, indices: Array) -> Array:
    """For each key, pick the k-th occurrence (relative to absolute index).

    Args:
        indices: Integer array of shape ``(len(keys),)`` with per-key
            relative indices. Values are taken modulo the key count.

    Returns:
        Integer array of shape ``(len(keys),)`` with absolute positions.
        Keys with zero occurrences return ``len(self)`` (OOB sentinel).
    """
    label_counts = self.counts.data
    indices = indices % jnp.maximum(label_counts, 1)
    sorted_indices = jnp.argsort(self.indices, stable=True)
    offsets = jnp.cumulative_sum(label_counts, include_initial=True)[:-1]
    result = sorted_indices.at[offsets + indices].get(
        mode="fill", fill_value=len(self)
    )
    return jnp.where(label_counts == 0, len(self), result)

subselect(needle, /, capacity, is_sorted=False)

Find elements matching target keys, returning key-level indices.

Parameters:

Name Type Description Default
needle Index[Key]

Index with target keys (must be a subset of self.keys).

required
capacity Capacity[int]

Capacity controlling output buffer size.

required
is_sorted bool

Whether self.indices is already sorted by key.

False

Returns:

Type Description
IndexSubselectResult[Key]

class:IndexSubselectResult with scatter/gather as Index[Key].

Source code in src/kups/core/data/index.py
def subselect(
    self, needle: Index[Key], /, capacity: Capacity[int], is_sorted: bool = False
) -> IndexSubselectResult[Key]:
    """Find elements matching target keys, returning key-level indices.

    Args:
        needle: Index with target keys (must be a subset of ``self.keys``).
        capacity: Capacity controlling output buffer size.
        is_sorted: Whether ``self.indices`` is already sorted by key.

    Returns:
        :class:`IndexSubselectResult` with scatter/gather as Index[Key].
    """
    result = subselect(
        needle.indices_in(self.keys),
        self.indices,
        output_buffer_size=capacity,
        num_segments=len(self.keys),
        is_sorted=is_sorted,
    )
    return IndexSubselectResult(
        scatter=Index(
            needle.keys,
            needle.indices.at[result.scatter_idxs].get(**needle.scatter_args),
            _cls=needle.cls,
        ),
        gather=Index(
            self.keys,
            self.indices.at[result.gather_idxs].get(**self.scatter_args),
            _cls=self.cls,
        ),
    )

sum_over(array)

Sums array values grouped by this index via segment sum.

Parameters:

Name Type Description Default
array Array

Array with leading dimension matching self.indices.

required

Returns:

Type Description
Table[Key, Array]

A Table mapping each key to its summed values.

Source code in src/kups/core/data/index.py
def sum_over(self, array: Array) -> Table[Key, Array]:
    """Sums ``array`` values grouped by this index via segment sum.

    Args:
        array: Array with leading dimension matching ``self.indices``.

    Returns:
        A ``Table`` mapping each key to its summed values.
    """
    from kups.core.data.table import Table

    return Table(
        self.keys,
        jax.ops.segment_sum(array, self.indices, self.num_labels, mode="drop"),
        _cls=self._cls,
    )

to_cls(new)

Convert keys to a different sentinel type.

Parameters:

Name Type Description Default
new Callable[[int], T2] | type[T2]

Type or callable mapping each integer key to the new type.

required

Returns:

Type Description
Index[T2]

New Index with converted keys, same indices and max_count.

Source code in src/kups/core/data/index.py
def to_cls[T: int, T2: SupportsSorting](
    self: Index[T], new: Callable[[int], T2] | type[T2]
) -> Index[T2]:
    """Convert keys to a different sentinel type.

    Args:
        new: Type or callable mapping each integer key to the new type.

    Returns:
        New ``Index`` with converted keys, same indices and max_count.
    """
    assert len(self.keys) > 0 or isinstance(new, type)
    new_keys = tuple(map(new, self.keys))
    cls = new if isinstance(new, type) else None
    return Index(new_keys, self.indices, max_count=self.max_count, _cls=cls)

transpose(*axes)

Transpose the index array, preserving keys.

Source code in src/kups/core/data/index.py
def transpose(self, *axes: int) -> Index[Key]:
    """Transpose the index array, preserving keys."""
    return self._forward_to_data("transpose", *axes)

update_labels(labels, *, allow_missing=False)

Re-index into a new key tuple, preserving logical mapping.

Parameters:

Name Type Description Default
labels tuple[Key, ...]

New key tuple (must be a superset of self.keys).

required
allow_missing bool

If True, keys in self that are absent from labels are mapped to the OOB sentinel instead of raising.

False

Returns:

Type Description
Index[Key]

New Index with labels and remapped indices.

Source code in src/kups/core/data/index.py
def update_labels(
    self, labels: tuple[Key, ...], *, allow_missing: bool = False
) -> Index[Key]:
    """Re-index into a new key tuple, preserving logical mapping.

    Args:
        labels: New key tuple (must be a superset of ``self.keys``).
        allow_missing: If ``True``, keys in ``self`` that are absent from
            ``labels`` are mapped to the OOB sentinel instead of raising.

    Returns:
        New ``Index`` with ``labels`` and remapped indices.
    """
    return Index(
        labels,
        self.indices_in(labels, allow_missing=allow_missing),
        max_count=self.max_count,
        _cls=self.cls,
    )

where_flat(target, /, capacity=None)

Find all positions matching any target key, flat.

Parameters:

Name Type Description Default
target Index[Key]

Index with keys to search for (must be a subset of self.keys).

required
capacity Capacity[int] | None

Capacity controlling output buffer size.

None

Returns:

Type Description
Array

Integer array of matching positions. Excess entries are filled

Array

with len(self) (OOB sentinel).

Source code in src/kups/core/data/index.py
def where_flat(
    self, target: Index[Key], /, capacity: Capacity[int] | None = None
) -> Array:
    """Find all positions matching any target key, flat.

    Args:
        target: Index with keys to search for (must be a subset of ``self.keys``).
        capacity: Capacity controlling output buffer size.

    Returns:
        Integer array of matching positions. Excess entries are filled
        with ``len(self)`` (OOB sentinel).
    """
    target_ids = target.indices_in(self.keys)
    mask = isin(self.indices, target_ids, len(self.keys))
    required = mask.sum()
    if is_traced(required):
        assert capacity is not None, "Capacity must be set during tracing."
        size = capacity.generate_assertion(required).size
    else:
        size = None
    return jnp.where(mask, size=size, fill_value=len(self))[0]

where_rectangular(target, max_count)

For each target key, find all positions padded to max_count.

Parameters:

Name Type Description Default
target Index[Key]

Index with keys to search for (must be a subset of self.keys).

required
max_count int

Maximum number of positions per key.

required

Returns:

Type Description
Array

Integer array of shape (len(target), max_count) with positions

Array

for each target key. Excess entries filled with len(self) (OOB).

Source code in src/kups/core/data/index.py
def where_rectangular(self, target: Index[Key], max_count: int) -> Array:
    """For each target key, find all positions padded to ``max_count``.

    Args:
        target: Index with keys to search for (must be a subset of ``self.keys``).
        max_count: Maximum number of positions per key.

    Returns:
        Integer array of shape ``(len(target), max_count)`` with positions
        for each target key. Excess entries filled with ``len(self)`` (OOB).
    """
    target_ids = target.indices_in(self.keys)

    @jax.vmap
    def _find(label: Array) -> Array:
        return jnp.where(
            self.indices == label, size=max_count, fill_value=len(self)
        )[0]

    return _find(target_ids)

zeros(shape, *, label=int, max_count=None) classmethod

Create an Index filled with a single key (the zeroth).

Parameters:

Name Type Description Default
shape int | tuple[int, ...]

Shape of the resulting index array.

required
label Callable[[int], T]

Callable mapping int to key values (default: int).

int
max_count int | None

Optional upper bound on occurrences per key.

None
Source code in src/kups/core/data/index.py
@classmethod
def zeros[T: SupportsSorting](
    cls,
    shape: int | tuple[int, ...],
    *,
    label: Callable[[int], T] = int,
    max_count: int | None = None,
) -> Index[T]:
    """Create an ``Index`` filled with a single key (the zeroth).

    Args:
        shape: Shape of the resulting index array.
        label: Callable mapping ``int`` to key values (default: ``int``).
        max_count: Optional upper bound on occurrences per key.
    """
    return cls.integer(
        np.zeros(shape, dtype=int), n=1, label=label, max_count=max_count
    )

IndexSubselectResult

Key-level scatter/gather result from :meth:Index.subselect.

For each match, scatter gives the matched key from the needle side and gather gives the matched key from the self (haystack) side. For valid entries both carry the same key value, encoded in their respective key spaces.

Attributes:

Name Type Description
scatter Index[Key]

Index[Key] -- matched key per entry from needle.

gather Index[Key]

Index[Key] -- matched key per entry from self.

Source code in src/kups/core/data/index.py
@dataclass
class IndexSubselectResult[Key: SupportsSorting]:
    """Key-level scatter/gather result from :meth:`Index.subselect`.

    For each match, ``scatter`` gives the matched key from the needle side
    and ``gather`` gives the matched key from the self (haystack) side.
    For valid entries both carry the same key value, encoded in their
    respective key spaces.

    Attributes:
        scatter: Index[Key] -- matched key per entry from needle.
        gather: Index[Key] -- matched key per entry from self.
    """

    scatter: Index[Key]
    gather: Index[Key]

    def __iter__(self):
        yield self.scatter
        yield self.gather

Sliceable

Bases: Batched

Batched dataclass with .at slicing and __getitem__ support.

Provides lens-based .at(index) for get/set and direct indexing via self[index].

Source code in src/kups/core/data/batched.py
class Sliceable(Batched):
    """Batched dataclass with ``.at`` slicing and ``__getitem__`` support.

    Provides lens-based ``.at(index)`` for get/set and direct indexing
    via ``self[index]``.
    """

    def at[S](self: S, index, **kwargs) -> BoundLens[S, S]:
        """Return a bound lens focused on ``index``."""
        return bind(self).at(index, **kwargs)

    def __getitem__(self, index) -> Self:
        """Gather entries at ``index``."""
        return self.at(index).get()

__getitem__(index)

Gather entries at index.

Source code in src/kups/core/data/batched.py
def __getitem__(self, index) -> Self:
    """Gather entries at ``index``."""
    return self.at(index).get()

at(index, **kwargs)

Return a bound lens focused on index.

Source code in src/kups/core/data/batched.py
def at[S](self: S, index, **kwargs) -> BoundLens[S, S]:
    """Return a bound lens focused on ``index``."""
    return bind(self).at(index, **kwargs)

Table

Bases: Batched, Generic[TKey, TData]

Entity-relation table with a primary key column and a data column.

A Table[TKey, TData] is a keyed data container analogous to a database table. keys is the primary key column (unique, sorted) and data is the value column — a pytree of arrays whose leaves share a leading dimension equal to len(keys).

Accessing data:

  • .data gives the raw value pytree, aligned to this table's own keys. Use for operations within a single key space.
  • table[index] where index: Index[TKey] performs a foreign-key lookup — gathering rows by key, analogous to a SQL JOIN. Use this whenever data must be broadcast across key spaces::

    # system → particle: broadcast system data to each particle per_particle = systems[particles.data.system]

    # particle → edge: broadcast particle data to each edge per_edge = particles[edges.indices]

Any Index[TKey] leaf inside another table's data acts as a foreign key referencing this table's primary keys.

Attributes:

Name Type Description
keys tuple[TKey, ...]

Primary key column — unique sorted tuple, one entry per row.

data TData

Value column — pytree of arrays with leading dimension len(keys).

Source code in src/kups/core/data/table.py
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
@dataclass
class Table(Batched, Generic[TKey, TData]):
    """Entity-relation table with a primary key column and a data column.

    A ``Table[TKey, TData]`` is a keyed data container analogous to a database
    table.  ``keys`` is the primary key column (unique, sorted) and ``data``
    is the value column — a pytree of arrays whose leaves share a leading
    dimension equal to ``len(keys)``.

    **Accessing data:**

    - ``.data`` gives the raw value pytree, aligned to this table's own keys.
      Use for operations within a single key space.
    - ``table[index]`` where ``index: Index[TKey]`` performs a foreign-key
      lookup — gathering rows by key, analogous to a SQL JOIN.  Use this
      whenever data must be broadcast across key spaces::

          # system → particle: broadcast system data to each particle
          per_particle = systems[particles.data.system]

          # particle → edge: broadcast particle data to each edge
          per_edge = particles[edges.indices]

      Any ``Index[TKey]`` leaf inside another table's ``data`` acts as a
      foreign key referencing this table's primary keys.

    Attributes:
        keys: Primary key column — unique sorted tuple, one entry per row.
        data: Value column — pytree of arrays with leading dimension ``len(keys)``.
    """

    keys: tuple[TKey, ...] = field(static=True)
    data: TData
    _cls: type[TKey] | None = field(static=True, default=None, kw_only=True)

    @property
    def cls(self) -> type[TKey]:
        """Key type, always available even for empty tables."""
        if self._cls is not None:
            return self._cls
        if len(self.keys) > 0:
            return type(self.keys[0])
        raise ValueError("Key type cannot be inferred from empty Table.")

    @skip_post_init_if_disabled
    def __post_init__(self):
        n = len(self.keys)
        assert len(np.unique(np.asarray(self.keys, dtype=object))) == n, (
            "Primary keys must be unique"
        )

        def _validate_and_broadcast(leaf):
            if isinstance(leaf, (Index, Table)):
                return leaf
            dim = leaf.shape[0]
            if dim != n and dim != 1:
                raise ValueError(
                    f"Leaf has size {dim} along axis 0, "
                    f"expected {n} (or 1 for broadcasting)."
                )
            if dim == 1:
                shape = (n,) + leaf.shape[1:]
                return jnp.broadcast_to(leaf, shape)
            return leaf

        updated_data = tree_map(
            _validate_and_broadcast,
            self.data,
            is_leaf=lambda x: isinstance(x, (Index, Table)),
        )
        assert all(type(x) is self.cls for x in self.keys), (
            f"Key type mismatch: expected {self.cls.__name__}, "
            f"got {set(type(x).__name__ for x in self.keys)}."
        )
        object.__setattr__(self, "_cls", self.cls)
        object.__setattr__(self, "data", updated_data)

    def __str__(self) -> str:
        keys_str = _format_keys(self.keys)
        return f"Table({self.cls.__name__}, keys={keys_str}, data={self.data})"

    @classmethod
    def arange[D, Label: SupportsSorting](
        cls, data: D, *, label: type[Label] = int
    ) -> Table[Label, D]:
        """Create a ``Table`` with keys ``label(0), label(1), ..., label(n-1)``."""
        leaves = jax.tree.leaves(data)
        n_items = leaves[0].shape[0]
        index = tuple(map(label, range(n_items)))
        return Table(index, data, _cls=label)

    def at[L: SupportsSorting, D](
        self: Table[L, D], index: Index[L], *, args: ScatterArgs | None = None
    ) -> BoundLens[D, D]:
        """Return a bound lens focused on entries selected by ``index``."""
        return bind(self.data).at(index.indices_in(self.keys), args=args)

    def update[D](
        self: Table[TKey, D], index: Index[TKey], data: D, **kwargs
    ) -> Table[TKey, D]:
        """Write ``data`` into rows selected by ``index``."""
        return (
            bind(self, lambda x: x.data)
            .at(index.indices_in(self.keys), **kwargs)
            .set(data)
        )

    def subset(self, index: Index[TKey]) -> Self:
        """Extract a subset of rows, re-keying as ``(0, 1, ...)``.

        Args:
            index: Rows to extract (must reference ``self.keys``).

        Returns:
            New container with freshly numbered keys.
        """
        new_data = self[index]
        new_idx = tuple(map(self.cls, range(len(index))))
        with no_post_init():
            result = bind(self, lambda x: (x.keys, x.data)).set((new_idx, new_data))
        return result

    def update_if[D, L: SupportsSorting](
        self: Table[TKey, D],
        accept: Table[L, Array],
        indices: Index[TKey],
        new_data: D,
    ) -> Table[TKey, D]:
        """Conditionally update rows based on a per-element accept mask.

        The accept mask is resolved against the indices found in both
        ``self[indices]`` and ``new_data``, and the union of the two is
        used to select which entries to write.

        Args:
            accept: Per-key boolean acceptance indexed by ``L``.
            indices: Target slot positions in ``self``.
            new_data: Proposed replacement data (same structure as subset).

        Returns:
            Updated container with accepted entries written.
        """
        target_cls = accept.cls
        current_data = self[indices]
        self_idx = Index.find(current_data, target_cls)
        data_idx = Index.find(new_data, target_cls)
        mask = (
            accept.at(self_idx, args={"mode": "fill", "fill_value": False}).get()
            | accept.at(data_idx, args={"mode": "fill", "fill_value": False}).get()
        )
        to_write = tree_where_broadcast_last(mask, new_data, current_data)
        return self.update(indices, to_write)

    @overload
    def __getitem__[L1: SupportsSorting, L2: SupportsSorting, L3: SupportsSorting, D](
        self: Table[TKey, Table[L1, Table[L2, Table[L3, D]]]],
        index: tuple[Index[TKey], Index[L1], Index[L2], Index[L3]],
    ) -> D: ...
    @overload
    def __getitem__[L1: SupportsSorting, L2: SupportsSorting, D](
        self: Table[TKey, Table[L1, Table[L2, D]]],
        index: tuple[Index[TKey], Index[L1], Index[L2]],
    ) -> D: ...
    @overload
    def __getitem__[L: SupportsSorting, D](
        self: Table[TKey, Table[L, D]], index: tuple[Index[TKey], Index[L]]
    ) -> D: ...
    @overload
    def __getitem__(self, index: Index[TKey]) -> TData: ...
    @no_type_check
    def __getitem__(self, index):
        """Retrieve data entries selected by ``index``."""
        if isinstance(index, tuple):
            result = self
            for idx in index:
                result = result[idx]
            return result
        idx = index.indices_in(self.keys)
        return bind(self.data).at((idx,)).get()

    def __len__(self) -> int:
        """Number of entries along the leading axis."""
        return len(self.keys)

    @property
    def size(self) -> int:
        """Number of entries along the leading axis (same as ``len()``)."""
        return len(self)

    def __contains__[L: SupportsSorting](self: Table[L, TData], key: L) -> bool:
        """Check whether ``key`` exists in the table."""
        return key in self.keys

    def map_data[L: SupportsSorting, D, T](
        self: Table[L, D], fn: Callable[[D], T]
    ) -> Table[L, T]:
        """Apply ``fn`` to ``data``, keeping the same keys."""
        return Table(self.keys, fn(self.data), _cls=self._cls)

    def set_data[L: SupportsSorting, D, T](self: Table[L, D], data: T) -> Table[L, T]:
        """Replace ``data``, keeping the same keys."""
        return Table(self.keys, data, _cls=self._cls)

    @staticmethod
    @overload
    def join[L: SupportsSorting, D, T1](
        base: Table[L, D], other: Table[L, T1], /
    ) -> Table[L, tuple[D, T1]]: ...
    @staticmethod
    @overload
    def join[L: SupportsSorting, D, T1, T2](
        base: Table[L, D], o1: Table[L, T1], o2: Table[L, T2], /
    ) -> Table[L, tuple[D, T1, T2]]: ...
    @staticmethod
    @overload
    def join[L: SupportsSorting, D, T1, T2, T3](
        base: Table[L, D],
        o1: Table[L, T1],
        o2: Table[L, T2],
        o3: Table[L, T3],
        /,
    ) -> Table[L, tuple[D, T1, T2, T3]]: ...
    @staticmethod
    @overload
    def join[L: SupportsSorting, D, T1, T2, T3, T4](
        base: Table[L, D],
        o1: Table[L, T1],
        o2: Table[L, T2],
        o3: Table[L, T3],
        o4: Table[L, T4],
        /,
    ) -> Table[L, tuple[D, T1, T2, T3, T4]]: ...
    @staticmethod
    @overload
    def join[L: SupportsSorting, D, T1, T2, T3, T4, T5](
        base: Table[L, D],
        o1: Table[L, T1],
        o2: Table[L, T2],
        o3: Table[L, T3],
        o4: Table[L, T4],
        o5: Table[L, T5],
        /,
    ) -> Table[L, tuple[D, T1, T2, T3, T4, T5]]: ...
    @staticmethod
    def join(base: Table, *others: Table) -> Table:
        """Join multiple ``Table`` objects on matching keys into tuple data.

        Performs a SQL-style ``JOIN`` on key equality. All arguments must
        share the same key set. If keys appear in a different order, the
        ``others`` are reindexed to match ``base``'s key ordering before
        their data is combined.

        Args:
            base: The reference ``Table`` whose key ordering is preserved.
            *others: One or more additional ``Table`` objects to join.
                Must have exactly the same key set as ``base``.

        Returns:
            A new ``Table`` with the same keys as ``base`` and data
            ``(base.data, others[0].data, ...)``.

        Raises:
            ValueError: If fewer than one ``other`` is provided, or if
                any argument's key set differs from ``base``.

        Example::

            >>> species = Table(("H", "O"), jnp.array([1, 8]))
            >>> masses  = Table(("O", "H"), jnp.array([16.0, 1.0]))
            >>> joined  = Table.join(species, masses)
            >>> joined.data  # (array([1, 8]), array([1.0, 16.0]))
        """
        if not others:
            raise ValueError("merge_tables requires at least two arguments")
        base, *others = Table.broadcast(base, *others)  # type: ignore

        def _align(other: Table, i: int):
            if other.keys == base.keys:
                return other.data
            if set(other.keys) != set(base.keys):
                raise ValueError(
                    f"Key set mismatch at argument {i + 1}: "
                    f"expected {set(base.keys)}, got {set(other.keys)}"
                )
            idx = Index.new(list(base.keys))
            return other[idx]

        aligned = [_align(o, i) for i, o in enumerate(others)]
        return Table(base.keys, (base.data, *aligned), _cls=base._cls)

    def __iter__(self):
        def data_slice(i: int) -> TData:
            return tree_map(lambda x: x[i], self.data)

        yield from ((key, data_slice(i)) for i, key in enumerate(self.keys))

    def slice(
        self, start: int = 0, end: int | None = None, step: int = 1
    ) -> Table[TKey, TData]:
        """Slice along the leading axis, preserving the corresponding keys.

        Args:
            start: Start index (default 0).
            end: End index (default ``None`` = end of array).
            step: Step size (default 1).
        """
        s = slice(start, end, step)
        data = tree_map(lambda x: x[s], self.data)
        index = self.keys[s]
        return Table(index, data, _cls=self.cls)

    @staticmethod
    def broadcast_to[L: SupportsSorting, D](
        source: Table[L, D], target: Table[L, Any]
    ) -> Table[L, D]:
        """Broadcast ``source`` to match the size of ``target``.

        Convenience wrapper around ``Table.broadcast(source, target)[0]``.
        ``source`` must either already have the same length as ``target``
        or have length 1 (in which case it is repeated).

        Args:
            source: The ``Table`` to broadcast.
            target: The ``Table`` whose size is matched.

        Returns:
            A ``Table`` with the same data as ``source``, broadcast to
            ``len(target)`` entries.
        """
        return Table.broadcast(source, target)[0]

    def __tree_match__(
        self, *others: Table[TKey, TData]
    ) -> tuple[Table[TKey, TData], ...]:
        return Table.broadcast(self, *others)

    @overload
    def __add__(
        self: Table[TKey, SupportsAdd[TData]], other: Table[TKey, TData]
    ) -> Table[TKey, TData]: ...
    @overload
    def __add__[D](
        self: Table[TKey, SupportsAdd[D]], other: D
    ) -> Table[TKey, TData]: ...
    def __add__(self, other) -> Table[TKey, TData]:
        return _table_operator(lambda a, b: a + b, self, other)

    @overload
    def __sub__(
        self: Table[TKey, SupportsSub[TData]], other: Table[TKey, TData]
    ) -> Table[TKey, TData]: ...
    @overload
    def __sub__[D](
        self: Table[TKey, SupportsSub[D]], other: D
    ) -> Table[TKey, TData]: ...
    def __sub__(self, other) -> Table[TKey, TData]:
        return _table_operator(lambda a, b: a - b, self, other)

    @overload
    def __mul__(
        self: Table[TKey, SupportsMul[TData]], other: Table[TKey, TData]
    ) -> Table[TKey, TData]: ...
    @overload
    def __mul__[D](
        self: Table[TKey, SupportsMul[D]], other: D
    ) -> Table[TKey, TData]: ...
    def __mul__(self, other) -> Table[TKey, TData]:
        return _table_operator(lambda a, b: a * b, self, other)

    @overload
    def __truediv__(
        self: Table[TKey, SupportsTrueDiv[TData]], other: Table[TKey, TData]
    ) -> Table[TKey, TData]: ...
    @overload
    def __truediv__[D](
        self: Table[TKey, SupportsTrueDiv[D]], other: D
    ) -> Table[TKey, TData]: ...
    def __truediv__(self, other) -> Table[TKey, TData]:
        return _table_operator(lambda a, b: a / b, self, other)

    @overload
    def __floordiv__(
        self: Table[TKey, SupportsFloorDiv[TData]], other: Table[TKey, TData]
    ) -> Table[TKey, TData]: ...
    @overload
    def __floordiv__[D](
        self: Table[TKey, SupportsFloorDiv[D]], other: D
    ) -> Table[TKey, TData]: ...
    def __floordiv__(self, other) -> Table[TKey, TData]:
        return _table_operator(lambda a, b: a // b, self, other)

    @overload
    def __gt__(
        self: Table[TKey, SupportsGt[TData]], other: Table[TKey, TData]
    ) -> Table[TKey, TData]: ...
    @overload
    def __gt__[D](self: Table[TKey, SupportsGt[D]], other: D) -> Table[TKey, TData]: ...
    def __gt__(self, other) -> Table[TKey, TData]:
        return _table_operator(lambda a, b: a > b, self, other)

    @staticmethod
    @overload
    def broadcast[L: SupportsSorting, D1](
        item1: Table[L, D1], /
    ) -> tuple[Table[L, D1]]: ...
    @staticmethod
    @overload
    def broadcast[L: SupportsSorting, D1, D2](
        item1: Table[L, D1], item2: Table[L, D2], /
    ) -> tuple[Table[L, D1], Table[L, D2]]: ...
    @staticmethod
    @overload
    def broadcast[L: SupportsSorting, D1, D2, D3](
        item1: Table[L, D1], item2: Table[L, D2], item3: Table[L, D3], /
    ) -> tuple[Table[L, D1], Table[L, D2], Table[L, D3]]: ...
    @staticmethod
    @overload
    def broadcast[L: SupportsSorting, D1, D2, D3, D4](
        item1: Table[L, D1],
        item2: Table[L, D2],
        item3: Table[L, D3],
        item4: Table[L, D4],
        /,
    ) -> tuple[Table[L, D1], Table[L, D2], Table[L, D3], Table[L, D4]]: ...
    @staticmethod
    @overload
    def broadcast[L: SupportsSorting](
        *items: Table[L, Any],
    ) -> tuple[Table[L, Any], ...]: ...
    @staticmethod
    def broadcast(
        *items: Table,
    ) -> tuple[Table, ...]:
        """Broadcast ``Table`` containers to a common leading-axis size.

        Analogous to NumPy broadcasting: all inputs must share the same
        key type, and each must either have the maximum size among inputs
        or have exactly size 1. Size-1 tables are expanded by repeating
        their single entry along the leading axis. Requires integer-based
        keys (e.g. ``SystemId``) so that the expanded key range
        ``0 .. max_size-1`` can be generated.

        Args:
            *items: One or more ``Table`` objects to broadcast. All must
                share the same key type. Each must have length equal to
                the maximum or length 1.

        Returns:
            A tuple of ``Table`` objects (one per input), all with the
            same leading-axis size and ``arange`` keys.

        Raises:
            AssertionError: If key types differ, sizes are not
                broadcastable, or keys of full-size tables are not
                ``arange``-style.

        Example::

            >>> scalars = Table.arange(jnp.array([1.0]), label=SystemId)
            >>> vectors = Table.arange(jnp.array([1, 2, 3]), label=SystemId)
            >>> s, v = Table.broadcast(scalars, vectors)
            >>> len(s)  # 3, was broadcast from 1
        """
        cls = items[0].cls
        assert all(cls is i.cls for i in items), (
            f"Key type mismatch: expected all {cls.__name__}, "
            f"got {[i.cls.__name__ for i in items if i.cls is not cls]}"
        )
        max_size = max(len(i) for i in items)

        if all(len(i) == max_size for i in items):
            return items

        sizes = [len(i) for i in items]
        assert all(s in (max_size, 1) for s in sizes), (
            f"Cannot broadcast Table sizes {sizes}: all must be {max_size} or 1"
        )
        assert issubclass(cls, int), (
            f"Broadcasting requires int-based keys, got {cls.__name__}"
        )
        assert all(
            len(i) == 1 or i.keys == tuple(map(cls, range(max_size))) for i in items
        ), (
            f"All full-size Table objects must have arange keys (0..{max_size - 1}) for broadcast"
        )

        new_index = tuple(map(cls, range(max_size)))
        result = [
            bind(i, lambda x: (x.keys, x.data)).set(
                (
                    new_index,
                    tree_map(
                        lambda y: jnp.broadcast_to(y, (max_size, *y.shape[1:])), i.data
                    ),
                )
            )
            for i in items
        ]
        return tuple(result)

    @staticmethod
    @overload
    def union[L1: SupportsSorting, D1](
        item1: Sequence[Table[L1, D1]],
        /,
    ) -> Table[L1, D1]: ...
    @staticmethod
    @overload
    def union[L1: SupportsSorting, D1, L2: SupportsSorting, D2](
        item1: Sequence[Table[L1, D1]],
        item2: Sequence[Table[L2, D2]],
        /,
    ) -> tuple[Table[L1, D1], Table[L2, D2]]: ...
    @staticmethod
    @overload
    def union[
        L1: SupportsSorting,
        D1,
        L2: SupportsSorting,
        D2,
        L3: SupportsSorting,
        D3,
    ](
        item1: Sequence[Table[L1, D1]],
        item2: Sequence[Table[L2, D2]],
        item3: Sequence[Table[L3, D3]],
        /,
    ) -> tuple[Table[L1, D1], Table[L2, D2], Table[L3, D3]]: ...
    @staticmethod
    @overload
    def union[
        L1: SupportsSorting,
        D1,
        L2: SupportsSorting,
        D2,
        L3: SupportsSorting,
        D3,
        L4: SupportsSorting,
        D4,
    ](
        item1: Sequence[Table[L1, D1]],
        item2: Sequence[Table[L2, D2]],
        item3: Sequence[Table[L3, D3]],
        item4: Sequence[Table[L4, D4]],
        /,
    ) -> tuple[Table[L1, D1], Table[L2, D2], Table[L3, D3], Table[L4, D4]]: ...
    @staticmethod
    def union(*groups: Sequence[Table]) -> tuple[Table, ...] | Table:
        """Concatenate multiple ``Table`` sequences (SQL ``UNION ALL``).

        Each positional argument is a sequence of ``Table`` objects that
        share the same key type and schema. Tables within each sequence
        are concatenated along the leading axis. Integer-based sentinel
        keys (e.g. ``SystemId``, ``ParticleId``) are offset-shifted per
        source so that the resulting keys are globally unique. Leaf
        ``Index`` objects nested inside ``data`` are similarly remapped.

        When multiple groups are given, leaf ``Index`` keys are first
        aligned across corresponding tables via ``Table.match`` so that
        cross-references (e.g. particles pointing at system ids) remain
        consistent after concatenation.

        Args:
            *groups: One or more sequences of ``Table`` objects. All
                sequences must have the same length (i.e. the same
                number of sources to merge). Each sequence represents
                one "column" of the database being unioned.

        Returns:
            A single ``Table`` if one group is provided, otherwise a
            tuple of ``Table`` objects (one per group).

        Raises:
            AssertionError: If group lengths differ or duplicate key
                types appear across groups.

        Example::

            >>> p0 = Table.arange(jnp.array([1, 8]), label=ParticleId)
            >>> p1 = Table.arange(jnp.array([6, 7]), label=ParticleId)
            >>> merged = Table.union([p0, p1])
            >>> len(merged)  # 4
            >>> merged.keys  # (ParticleId(0), ..., ParticleId(3))
        """
        n = len(groups[0])
        assert all(len(g) == n for g in groups), "All groups must have the same length"
        groups = tuple(zip(*map(lambda x: Table.match(*x), zip(*groups))))

        def _key_type(group: Sequence[Table]) -> type | None:
            for item in group:
                if item._cls is not None:
                    return item._cls
                if item.keys:
                    return type(item.keys[0])
            return None

        offsets: dict[type, list[int]] = {}
        for group in groups:
            lt = _key_type(group)
            if lt is None:
                continue
            assert lt not in offsets, f"Duplicate key type {lt.__name__} across groups"
            acc = 0
            offsets[lt] = []
            for item in group:
                offsets[lt].append(acc)
                acc += len(item)

        def _shift_key(key, off: int):
            return type(key)(int(key) + off) if isinstance(key, int) else key

        def _concat_leaves(*leaves):
            if isinstance(leaves[0], Index):
                if leaves[0].cls in offsets:
                    return Index.concatenate(*leaves, shift_keys=True)
                return Index.concatenate(*leaves)
            return jnp.concatenate(leaves, axis=0)

        is_index = lambda x: isinstance(x, Index)  # noqa: E731
        results: list[Table] = []
        for group in groups:
            lt = _key_type(group)
            if lt is None:
                results.append(group[0])
                continue
            merged_index = tuple(
                _shift_key(key, offsets[lt][i])
                for i, item in enumerate(group)
                for key in item.keys
            )
            merged_data = jax.tree.map(
                _concat_leaves, *(item.data for item in group), is_leaf=is_index
            )
            results.append(Table(merged_index, merged_data, _cls=lt))

        return results[0] if len(results) == 1 else tuple(results)

    @staticmethod
    @overload
    def match[L1: SupportsSorting, D1](
        group1: Table[L1, D1], /
    ) -> tuple[Table[L1, D1]]: ...
    @staticmethod
    @overload
    def match[L1: SupportsSorting, D1, L2: SupportsSorting, D2](
        group1: Table[L1, D1], group2: Table[L2, D2], /
    ) -> tuple[Table[L1, D1], Table[L2, D2]]: ...
    @staticmethod
    @overload
    def match[
        L1: SupportsSorting,
        D1,
        L2: SupportsSorting,
        D2,
        L3: SupportsSorting,
        D3,
    ](
        group1: Table[L1, D1],
        group2: Table[L2, D2],
        group3: Table[L3, D3],
        /,
    ) -> tuple[Table[L1, D1], Table[L2, D2], Table[L3, D3]]: ...
    @staticmethod
    @overload
    def match[
        L1: SupportsSorting,
        D1,
        L2: SupportsSorting,
        D2,
        L3: SupportsSorting,
        D3,
        L4: SupportsSorting,
        D4,
    ](
        group1: Table[L1, D1],
        group2: Table[L2, D2],
        group3: Table[L3, D3],
        group4: Table[L4, D4],
        /,
    ) -> tuple[Table[L1, D1], Table[L2, D2], Table[L3, D3], Table[L4, D4]]: ...
    @staticmethod
    def match(*groups: Table) -> tuple[Table, ...] | Table:
        """Align leaf ``Index`` keys across multiple ``Table`` containers.

        Ensures that all ``Index`` leaves of the same key type share an
        identical key vocabulary. For each key type present across the
        inputs, the key tuples are merged (deduplicated, sorted) and
        every ``Index`` leaf of that type is updated to use the shared
        vocabulary via ``Index.update_labels``.

        This is typically called before operations that require
        element-wise comparison of indices across tables (e.g. before
        ``Table.union``).

        Args:
            *groups: One or more ``Table`` objects whose leaf indices
                should be aligned.

        Returns:
            A tuple of ``Table`` objects with unified ``Index`` keys,
            or a single ``Table`` if only one input is given.
        """
        indices = {g.cls: g.keys for g in groups}

        def traversal(x):
            if not isinstance(x, Index):
                return x
            if x.cls in indices:
                return x.update_labels(indices[x.cls])
            return x

        result = tree_map(traversal, groups, is_leaf=lambda x: isinstance(x, Index))
        return result

    @staticmethod
    @overload
    def transform[L: SupportsSorting, D1, R](
        fn: Callable[[D1], R],
    ) -> Callable[[Table[L, D1]], Table[L, R]]: ...
    @staticmethod
    @overload
    def transform[L: SupportsSorting, D1, D2, R](
        fn: Callable[[D1, D2], R],
    ) -> Callable[[Table[L, D1], Table[L, D2]], Table[L, R]]: ...
    @staticmethod
    @overload
    def transform[L: SupportsSorting, D1, D2, D3, R](
        fn: Callable[[D1, D2, D3], R],
    ) -> Callable[[Table[L, D1], Table[L, D2], Table[L, D3]], Table[L, R]]: ...
    @staticmethod
    @overload
    def transform[L: SupportsSorting, D1, D2, D3, D4, R](
        fn: Callable[[D1, D2, D3, D4], R],
    ) -> Callable[
        [Table[L, D1], Table[L, D2], Table[L, D3], Table[L, D4]], Table[L, R]
    ]: ...
    @staticmethod
    def transform(fn: Callable[..., Any]) -> Callable[..., Any]:
        """Lift a function on raw data to operate on ``Table`` containers.

        Returns a wrapper that unpacks ``.data`` from each ``Table``
        argument, calls ``fn``, and re-wraps the result in a new
        ``Table`` with the same keys. All inputs must share identical
        keys.

        Args:
            fn: A callable ``(D1, D2, ...) -> R`` operating on the raw
                data payloads of the tables.

        Returns:
            A callable ``(Table[L, D1], Table[L, D2], ...) -> Table[L, R]``
            that applies ``fn`` element-wise over the data.

        Example::

            >>> double = Table.transform(lambda x: x * 2)
            >>> t = Table.arange(jnp.array([1, 2, 3]), label=SystemId)
            >>> double(t).data  # array([2, 4, 6])
        """

        def wrapped(*args: Table[Any, Any]) -> Table[Any, Any]:
            assert all(a.keys == args[0].keys for a in args)
            return Table(args[0].keys, fn(*[a.data for a in args]))

        return wrapped

cls property

Key type, always available even for empty tables.

size property

Number of entries along the leading axis (same as len()).

__contains__(key)

Check whether key exists in the table.

Source code in src/kups/core/data/table.py
def __contains__[L: SupportsSorting](self: Table[L, TData], key: L) -> bool:
    """Check whether ``key`` exists in the table."""
    return key in self.keys

__getitem__(index)

__getitem__(
    index: tuple[
        Index[TKey], Index[L1], Index[L2], Index[L3]
    ],
) -> D
__getitem__(
    index: tuple[Index[TKey], Index[L1], Index[L2]],
) -> D
__getitem__(index: tuple[Index[TKey], Index[L]]) -> D
__getitem__(index: Index[TKey]) -> TData

Retrieve data entries selected by index.

Source code in src/kups/core/data/table.py
@no_type_check
def __getitem__(self, index):
    """Retrieve data entries selected by ``index``."""
    if isinstance(index, tuple):
        result = self
        for idx in index:
            result = result[idx]
        return result
    idx = index.indices_in(self.keys)
    return bind(self.data).at((idx,)).get()

__len__()

Number of entries along the leading axis.

Source code in src/kups/core/data/table.py
def __len__(self) -> int:
    """Number of entries along the leading axis."""
    return len(self.keys)

arange(data, *, label=int) classmethod

Create a Table with keys label(0), label(1), ..., label(n-1).

Source code in src/kups/core/data/table.py
@classmethod
def arange[D, Label: SupportsSorting](
    cls, data: D, *, label: type[Label] = int
) -> Table[Label, D]:
    """Create a ``Table`` with keys ``label(0), label(1), ..., label(n-1)``."""
    leaves = jax.tree.leaves(data)
    n_items = leaves[0].shape[0]
    index = tuple(map(label, range(n_items)))
    return Table(index, data, _cls=label)

at(index, *, args=None)

Return a bound lens focused on entries selected by index.

Source code in src/kups/core/data/table.py
def at[L: SupportsSorting, D](
    self: Table[L, D], index: Index[L], *, args: ScatterArgs | None = None
) -> BoundLens[D, D]:
    """Return a bound lens focused on entries selected by ``index``."""
    return bind(self.data).at(index.indices_in(self.keys), args=args)

broadcast(*items) staticmethod

broadcast(item1: Table[L, D1]) -> tuple[Table[L, D1]]
broadcast(
    item1: Table[L, D1], item2: Table[L, D2]
) -> tuple[Table[L, D1], Table[L, D2]]
broadcast(
    item1: Table[L, D1],
    item2: Table[L, D2],
    item3: Table[L, D3],
) -> tuple[Table[L, D1], Table[L, D2], Table[L, D3]]
broadcast(
    item1: Table[L, D1],
    item2: Table[L, D2],
    item3: Table[L, D3],
    item4: Table[L, D4],
) -> tuple[
    Table[L, D1], Table[L, D2], Table[L, D3], Table[L, D4]
]
broadcast(
    *items: Table[L, Any],
) -> tuple[Table[L, Any], ...]

Broadcast Table containers to a common leading-axis size.

Analogous to NumPy broadcasting: all inputs must share the same key type, and each must either have the maximum size among inputs or have exactly size 1. Size-1 tables are expanded by repeating their single entry along the leading axis. Requires integer-based keys (e.g. SystemId) so that the expanded key range 0 .. max_size-1 can be generated.

Parameters:

Name Type Description Default
*items Table

One or more Table objects to broadcast. All must share the same key type. Each must have length equal to the maximum or length 1.

()

Returns:

Type Description
Table

A tuple of Table objects (one per input), all with the

...

same leading-axis size and arange keys.

Raises:

Type Description
AssertionError

If key types differ, sizes are not broadcastable, or keys of full-size tables are not arange-style.

Example::

>>> scalars = Table.arange(jnp.array([1.0]), label=SystemId)
>>> vectors = Table.arange(jnp.array([1, 2, 3]), label=SystemId)
>>> s, v = Table.broadcast(scalars, vectors)
>>> len(s)  # 3, was broadcast from 1
Source code in src/kups/core/data/table.py
@staticmethod
def broadcast(
    *items: Table,
) -> tuple[Table, ...]:
    """Broadcast ``Table`` containers to a common leading-axis size.

    Analogous to NumPy broadcasting: all inputs must share the same
    key type, and each must either have the maximum size among inputs
    or have exactly size 1. Size-1 tables are expanded by repeating
    their single entry along the leading axis. Requires integer-based
    keys (e.g. ``SystemId``) so that the expanded key range
    ``0 .. max_size-1`` can be generated.

    Args:
        *items: One or more ``Table`` objects to broadcast. All must
            share the same key type. Each must have length equal to
            the maximum or length 1.

    Returns:
        A tuple of ``Table`` objects (one per input), all with the
        same leading-axis size and ``arange`` keys.

    Raises:
        AssertionError: If key types differ, sizes are not
            broadcastable, or keys of full-size tables are not
            ``arange``-style.

    Example::

        >>> scalars = Table.arange(jnp.array([1.0]), label=SystemId)
        >>> vectors = Table.arange(jnp.array([1, 2, 3]), label=SystemId)
        >>> s, v = Table.broadcast(scalars, vectors)
        >>> len(s)  # 3, was broadcast from 1
    """
    cls = items[0].cls
    assert all(cls is i.cls for i in items), (
        f"Key type mismatch: expected all {cls.__name__}, "
        f"got {[i.cls.__name__ for i in items if i.cls is not cls]}"
    )
    max_size = max(len(i) for i in items)

    if all(len(i) == max_size for i in items):
        return items

    sizes = [len(i) for i in items]
    assert all(s in (max_size, 1) for s in sizes), (
        f"Cannot broadcast Table sizes {sizes}: all must be {max_size} or 1"
    )
    assert issubclass(cls, int), (
        f"Broadcasting requires int-based keys, got {cls.__name__}"
    )
    assert all(
        len(i) == 1 or i.keys == tuple(map(cls, range(max_size))) for i in items
    ), (
        f"All full-size Table objects must have arange keys (0..{max_size - 1}) for broadcast"
    )

    new_index = tuple(map(cls, range(max_size)))
    result = [
        bind(i, lambda x: (x.keys, x.data)).set(
            (
                new_index,
                tree_map(
                    lambda y: jnp.broadcast_to(y, (max_size, *y.shape[1:])), i.data
                ),
            )
        )
        for i in items
    ]
    return tuple(result)

broadcast_to(source, target) staticmethod

Broadcast source to match the size of target.

Convenience wrapper around Table.broadcast(source, target)[0]. source must either already have the same length as target or have length 1 (in which case it is repeated).

Parameters:

Name Type Description Default
source Table[L, D]

The Table to broadcast.

required
target Table[L, Any]

The Table whose size is matched.

required

Returns:

Type Description
Table[L, D]

A Table with the same data as source, broadcast to

Table[L, D]

len(target) entries.

Source code in src/kups/core/data/table.py
@staticmethod
def broadcast_to[L: SupportsSorting, D](
    source: Table[L, D], target: Table[L, Any]
) -> Table[L, D]:
    """Broadcast ``source`` to match the size of ``target``.

    Convenience wrapper around ``Table.broadcast(source, target)[0]``.
    ``source`` must either already have the same length as ``target``
    or have length 1 (in which case it is repeated).

    Args:
        source: The ``Table`` to broadcast.
        target: The ``Table`` whose size is matched.

    Returns:
        A ``Table`` with the same data as ``source``, broadcast to
        ``len(target)`` entries.
    """
    return Table.broadcast(source, target)[0]

join(base, *others) staticmethod

join(
    base: Table[L, D], other: Table[L, T1]
) -> Table[L, tuple[D, T1]]
join(
    base: Table[L, D], o1: Table[L, T1], o2: Table[L, T2]
) -> Table[L, tuple[D, T1, T2]]
join(
    base: Table[L, D],
    o1: Table[L, T1],
    o2: Table[L, T2],
    o3: Table[L, T3],
) -> Table[L, tuple[D, T1, T2, T3]]
join(
    base: Table[L, D],
    o1: Table[L, T1],
    o2: Table[L, T2],
    o3: Table[L, T3],
    o4: Table[L, T4],
) -> Table[L, tuple[D, T1, T2, T3, T4]]
join(
    base: Table[L, D],
    o1: Table[L, T1],
    o2: Table[L, T2],
    o3: Table[L, T3],
    o4: Table[L, T4],
    o5: Table[L, T5],
) -> Table[L, tuple[D, T1, T2, T3, T4, T5]]

Join multiple Table objects on matching keys into tuple data.

Performs a SQL-style JOIN on key equality. All arguments must share the same key set. If keys appear in a different order, the others are reindexed to match base's key ordering before their data is combined.

Parameters:

Name Type Description Default
base Table

The reference Table whose key ordering is preserved.

required
*others Table

One or more additional Table objects to join. Must have exactly the same key set as base.

()

Returns:

Type Description
Table

A new Table with the same keys as base and data

Table

(base.data, others[0].data, ...).

Raises:

Type Description
ValueError

If fewer than one other is provided, or if any argument's key set differs from base.

Example::

>>> species = Table(("H", "O"), jnp.array([1, 8]))
>>> masses  = Table(("O", "H"), jnp.array([16.0, 1.0]))
>>> joined  = Table.join(species, masses)
>>> joined.data  # (array([1, 8]), array([1.0, 16.0]))
Source code in src/kups/core/data/table.py
@staticmethod
def join(base: Table, *others: Table) -> Table:
    """Join multiple ``Table`` objects on matching keys into tuple data.

    Performs a SQL-style ``JOIN`` on key equality. All arguments must
    share the same key set. If keys appear in a different order, the
    ``others`` are reindexed to match ``base``'s key ordering before
    their data is combined.

    Args:
        base: The reference ``Table`` whose key ordering is preserved.
        *others: One or more additional ``Table`` objects to join.
            Must have exactly the same key set as ``base``.

    Returns:
        A new ``Table`` with the same keys as ``base`` and data
        ``(base.data, others[0].data, ...)``.

    Raises:
        ValueError: If fewer than one ``other`` is provided, or if
            any argument's key set differs from ``base``.

    Example::

        >>> species = Table(("H", "O"), jnp.array([1, 8]))
        >>> masses  = Table(("O", "H"), jnp.array([16.0, 1.0]))
        >>> joined  = Table.join(species, masses)
        >>> joined.data  # (array([1, 8]), array([1.0, 16.0]))
    """
    if not others:
        raise ValueError("merge_tables requires at least two arguments")
    base, *others = Table.broadcast(base, *others)  # type: ignore

    def _align(other: Table, i: int):
        if other.keys == base.keys:
            return other.data
        if set(other.keys) != set(base.keys):
            raise ValueError(
                f"Key set mismatch at argument {i + 1}: "
                f"expected {set(base.keys)}, got {set(other.keys)}"
            )
        idx = Index.new(list(base.keys))
        return other[idx]

    aligned = [_align(o, i) for i, o in enumerate(others)]
    return Table(base.keys, (base.data, *aligned), _cls=base._cls)

map_data(fn)

Apply fn to data, keeping the same keys.

Source code in src/kups/core/data/table.py
def map_data[L: SupportsSorting, D, T](
    self: Table[L, D], fn: Callable[[D], T]
) -> Table[L, T]:
    """Apply ``fn`` to ``data``, keeping the same keys."""
    return Table(self.keys, fn(self.data), _cls=self._cls)

match(*groups) staticmethod

match(group1: Table[L1, D1]) -> tuple[Table[L1, D1]]
match(
    group1: Table[L1, D1], group2: Table[L2, D2]
) -> tuple[Table[L1, D1], Table[L2, D2]]
match(
    group1: Table[L1, D1],
    group2: Table[L2, D2],
    group3: Table[L3, D3],
) -> tuple[Table[L1, D1], Table[L2, D2], Table[L3, D3]]
match(
    group1: Table[L1, D1],
    group2: Table[L2, D2],
    group3: Table[L3, D3],
    group4: Table[L4, D4],
) -> tuple[
    Table[L1, D1],
    Table[L2, D2],
    Table[L3, D3],
    Table[L4, D4],
]

Align leaf Index keys across multiple Table containers.

Ensures that all Index leaves of the same key type share an identical key vocabulary. For each key type present across the inputs, the key tuples are merged (deduplicated, sorted) and every Index leaf of that type is updated to use the shared vocabulary via Index.update_labels.

This is typically called before operations that require element-wise comparison of indices across tables (e.g. before Table.union).

Parameters:

Name Type Description Default
*groups Table

One or more Table objects whose leaf indices should be aligned.

()

Returns:

Type Description
tuple[Table, ...] | Table

A tuple of Table objects with unified Index keys,

tuple[Table, ...] | Table

or a single Table if only one input is given.

Source code in src/kups/core/data/table.py
@staticmethod
def match(*groups: Table) -> tuple[Table, ...] | Table:
    """Align leaf ``Index`` keys across multiple ``Table`` containers.

    Ensures that all ``Index`` leaves of the same key type share an
    identical key vocabulary. For each key type present across the
    inputs, the key tuples are merged (deduplicated, sorted) and
    every ``Index`` leaf of that type is updated to use the shared
    vocabulary via ``Index.update_labels``.

    This is typically called before operations that require
    element-wise comparison of indices across tables (e.g. before
    ``Table.union``).

    Args:
        *groups: One or more ``Table`` objects whose leaf indices
            should be aligned.

    Returns:
        A tuple of ``Table`` objects with unified ``Index`` keys,
        or a single ``Table`` if only one input is given.
    """
    indices = {g.cls: g.keys for g in groups}

    def traversal(x):
        if not isinstance(x, Index):
            return x
        if x.cls in indices:
            return x.update_labels(indices[x.cls])
        return x

    result = tree_map(traversal, groups, is_leaf=lambda x: isinstance(x, Index))
    return result

set_data(data)

Replace data, keeping the same keys.

Source code in src/kups/core/data/table.py
def set_data[L: SupportsSorting, D, T](self: Table[L, D], data: T) -> Table[L, T]:
    """Replace ``data``, keeping the same keys."""
    return Table(self.keys, data, _cls=self._cls)

slice(start=0, end=None, step=1)

Slice along the leading axis, preserving the corresponding keys.

Parameters:

Name Type Description Default
start int

Start index (default 0).

0
end int | None

End index (default None = end of array).

None
step int

Step size (default 1).

1
Source code in src/kups/core/data/table.py
def slice(
    self, start: int = 0, end: int | None = None, step: int = 1
) -> Table[TKey, TData]:
    """Slice along the leading axis, preserving the corresponding keys.

    Args:
        start: Start index (default 0).
        end: End index (default ``None`` = end of array).
        step: Step size (default 1).
    """
    s = slice(start, end, step)
    data = tree_map(lambda x: x[s], self.data)
    index = self.keys[s]
    return Table(index, data, _cls=self.cls)

subset(index)

Extract a subset of rows, re-keying as (0, 1, ...).

Parameters:

Name Type Description Default
index Index[TKey]

Rows to extract (must reference self.keys).

required

Returns:

Type Description
Self

New container with freshly numbered keys.

Source code in src/kups/core/data/table.py
def subset(self, index: Index[TKey]) -> Self:
    """Extract a subset of rows, re-keying as ``(0, 1, ...)``.

    Args:
        index: Rows to extract (must reference ``self.keys``).

    Returns:
        New container with freshly numbered keys.
    """
    new_data = self[index]
    new_idx = tuple(map(self.cls, range(len(index))))
    with no_post_init():
        result = bind(self, lambda x: (x.keys, x.data)).set((new_idx, new_data))
    return result

transform(fn) staticmethod

transform(
    fn: Callable[[D1], R],
) -> Callable[[Table[L, D1]], Table[L, R]]
transform(
    fn: Callable[[D1, D2], R],
) -> Callable[[Table[L, D1], Table[L, D2]], Table[L, R]]
transform(
    fn: Callable[[D1, D2, D3], R],
) -> Callable[
    [Table[L, D1], Table[L, D2], Table[L, D3]], Table[L, R]
]
transform(
    fn: Callable[[D1, D2, D3, D4], R],
) -> Callable[
    [
        Table[L, D1],
        Table[L, D2],
        Table[L, D3],
        Table[L, D4],
    ],
    Table[L, R],
]

Lift a function on raw data to operate on Table containers.

Returns a wrapper that unpacks .data from each Table argument, calls fn, and re-wraps the result in a new Table with the same keys. All inputs must share identical keys.

Parameters:

Name Type Description Default
fn Callable[..., Any]

A callable (D1, D2, ...) -> R operating on the raw data payloads of the tables.

required

Returns:

Type Description
Callable[..., Any]

A callable (Table[L, D1], Table[L, D2], ...) -> Table[L, R]

Callable[..., Any]

that applies fn element-wise over the data.

Example::

>>> double = Table.transform(lambda x: x * 2)
>>> t = Table.arange(jnp.array([1, 2, 3]), label=SystemId)
>>> double(t).data  # array([2, 4, 6])
Source code in src/kups/core/data/table.py
@staticmethod
def transform(fn: Callable[..., Any]) -> Callable[..., Any]:
    """Lift a function on raw data to operate on ``Table`` containers.

    Returns a wrapper that unpacks ``.data`` from each ``Table``
    argument, calls ``fn``, and re-wraps the result in a new
    ``Table`` with the same keys. All inputs must share identical
    keys.

    Args:
        fn: A callable ``(D1, D2, ...) -> R`` operating on the raw
            data payloads of the tables.

    Returns:
        A callable ``(Table[L, D1], Table[L, D2], ...) -> Table[L, R]``
        that applies ``fn`` element-wise over the data.

    Example::

        >>> double = Table.transform(lambda x: x * 2)
        >>> t = Table.arange(jnp.array([1, 2, 3]), label=SystemId)
        >>> double(t).data  # array([2, 4, 6])
    """

    def wrapped(*args: Table[Any, Any]) -> Table[Any, Any]:
        assert all(a.keys == args[0].keys for a in args)
        return Table(args[0].keys, fn(*[a.data for a in args]))

    return wrapped

union(*groups) staticmethod

union(item1: Sequence[Table[L1, D1]]) -> Table[L1, D1]
union(
    item1: Sequence[Table[L1, D1]],
    item2: Sequence[Table[L2, D2]],
) -> tuple[Table[L1, D1], Table[L2, D2]]
union(
    item1: Sequence[Table[L1, D1]],
    item2: Sequence[Table[L2, D2]],
    item3: Sequence[Table[L3, D3]],
) -> tuple[Table[L1, D1], Table[L2, D2], Table[L3, D3]]
union(
    item1: Sequence[Table[L1, D1]],
    item2: Sequence[Table[L2, D2]],
    item3: Sequence[Table[L3, D3]],
    item4: Sequence[Table[L4, D4]],
) -> tuple[
    Table[L1, D1],
    Table[L2, D2],
    Table[L3, D3],
    Table[L4, D4],
]

Concatenate multiple Table sequences (SQL UNION ALL).

Each positional argument is a sequence of Table objects that share the same key type and schema. Tables within each sequence are concatenated along the leading axis. Integer-based sentinel keys (e.g. SystemId, ParticleId) are offset-shifted per source so that the resulting keys are globally unique. Leaf Index objects nested inside data are similarly remapped.

When multiple groups are given, leaf Index keys are first aligned across corresponding tables via Table.match so that cross-references (e.g. particles pointing at system ids) remain consistent after concatenation.

Parameters:

Name Type Description Default
*groups Sequence[Table]

One or more sequences of Table objects. All sequences must have the same length (i.e. the same number of sources to merge). Each sequence represents one "column" of the database being unioned.

()

Returns:

Type Description
tuple[Table, ...] | Table

A single Table if one group is provided, otherwise a

tuple[Table, ...] | Table

tuple of Table objects (one per group).

Raises:

Type Description
AssertionError

If group lengths differ or duplicate key types appear across groups.

Example::

>>> p0 = Table.arange(jnp.array([1, 8]), label=ParticleId)
>>> p1 = Table.arange(jnp.array([6, 7]), label=ParticleId)
>>> merged = Table.union([p0, p1])
>>> len(merged)  # 4
>>> merged.keys  # (ParticleId(0), ..., ParticleId(3))
Source code in src/kups/core/data/table.py
@staticmethod
def union(*groups: Sequence[Table]) -> tuple[Table, ...] | Table:
    """Concatenate multiple ``Table`` sequences (SQL ``UNION ALL``).

    Each positional argument is a sequence of ``Table`` objects that
    share the same key type and schema. Tables within each sequence
    are concatenated along the leading axis. Integer-based sentinel
    keys (e.g. ``SystemId``, ``ParticleId``) are offset-shifted per
    source so that the resulting keys are globally unique. Leaf
    ``Index`` objects nested inside ``data`` are similarly remapped.

    When multiple groups are given, leaf ``Index`` keys are first
    aligned across corresponding tables via ``Table.match`` so that
    cross-references (e.g. particles pointing at system ids) remain
    consistent after concatenation.

    Args:
        *groups: One or more sequences of ``Table`` objects. All
            sequences must have the same length (i.e. the same
            number of sources to merge). Each sequence represents
            one "column" of the database being unioned.

    Returns:
        A single ``Table`` if one group is provided, otherwise a
        tuple of ``Table`` objects (one per group).

    Raises:
        AssertionError: If group lengths differ or duplicate key
            types appear across groups.

    Example::

        >>> p0 = Table.arange(jnp.array([1, 8]), label=ParticleId)
        >>> p1 = Table.arange(jnp.array([6, 7]), label=ParticleId)
        >>> merged = Table.union([p0, p1])
        >>> len(merged)  # 4
        >>> merged.keys  # (ParticleId(0), ..., ParticleId(3))
    """
    n = len(groups[0])
    assert all(len(g) == n for g in groups), "All groups must have the same length"
    groups = tuple(zip(*map(lambda x: Table.match(*x), zip(*groups))))

    def _key_type(group: Sequence[Table]) -> type | None:
        for item in group:
            if item._cls is not None:
                return item._cls
            if item.keys:
                return type(item.keys[0])
        return None

    offsets: dict[type, list[int]] = {}
    for group in groups:
        lt = _key_type(group)
        if lt is None:
            continue
        assert lt not in offsets, f"Duplicate key type {lt.__name__} across groups"
        acc = 0
        offsets[lt] = []
        for item in group:
            offsets[lt].append(acc)
            acc += len(item)

    def _shift_key(key, off: int):
        return type(key)(int(key) + off) if isinstance(key, int) else key

    def _concat_leaves(*leaves):
        if isinstance(leaves[0], Index):
            if leaves[0].cls in offsets:
                return Index.concatenate(*leaves, shift_keys=True)
            return Index.concatenate(*leaves)
        return jnp.concatenate(leaves, axis=0)

    is_index = lambda x: isinstance(x, Index)  # noqa: E731
    results: list[Table] = []
    for group in groups:
        lt = _key_type(group)
        if lt is None:
            results.append(group[0])
            continue
        merged_index = tuple(
            _shift_key(key, offsets[lt][i])
            for i, item in enumerate(group)
            for key in item.keys
        )
        merged_data = jax.tree.map(
            _concat_leaves, *(item.data for item in group), is_leaf=is_index
        )
        results.append(Table(merged_index, merged_data, _cls=lt))

    return results[0] if len(results) == 1 else tuple(results)

update(index, data, **kwargs)

Write data into rows selected by index.

Source code in src/kups/core/data/table.py
def update[D](
    self: Table[TKey, D], index: Index[TKey], data: D, **kwargs
) -> Table[TKey, D]:
    """Write ``data`` into rows selected by ``index``."""
    return (
        bind(self, lambda x: x.data)
        .at(index.indices_in(self.keys), **kwargs)
        .set(data)
    )

update_if(accept, indices, new_data)

Conditionally update rows based on a per-element accept mask.

The accept mask is resolved against the indices found in both self[indices] and new_data, and the union of the two is used to select which entries to write.

Parameters:

Name Type Description Default
accept Table[L, Array]

Per-key boolean acceptance indexed by L.

required
indices Index[TKey]

Target slot positions in self.

required
new_data D

Proposed replacement data (same structure as subset).

required

Returns:

Type Description
Table[TKey, D]

Updated container with accepted entries written.

Source code in src/kups/core/data/table.py
def update_if[D, L: SupportsSorting](
    self: Table[TKey, D],
    accept: Table[L, Array],
    indices: Index[TKey],
    new_data: D,
) -> Table[TKey, D]:
    """Conditionally update rows based on a per-element accept mask.

    The accept mask is resolved against the indices found in both
    ``self[indices]`` and ``new_data``, and the union of the two is
    used to select which entries to write.

    Args:
        accept: Per-key boolean acceptance indexed by ``L``.
        indices: Target slot positions in ``self``.
        new_data: Proposed replacement data (same structure as subset).

    Returns:
        Updated container with accepted entries written.
    """
    target_cls = accept.cls
    current_data = self[indices]
    self_idx = Index.find(current_data, target_cls)
    data_idx = Index.find(new_data, target_cls)
    mask = (
        accept.at(self_idx, args={"mode": "fill", "fill_value": False}).get()
        | accept.at(data_idx, args={"mode": "fill", "fill_value": False}).get()
    )
    to_write = tree_where_broadcast_last(mask, new_data, current_data)
    return self.update(indices, to_write)

WithCache

Bases: Generic[TData, TCache]

Data paired with an associated cache.

Attributes:

Name Type Description
data TData

Primary data.

cache TCache

Cached auxiliary values derived from or associated with data.

Source code in src/kups/core/data/wrappers.py
@dataclass
class WithCache(Generic[TData, TCache]):
    """Data paired with an associated cache.

    Attributes:
        data: Primary data.
        cache: Cached auxiliary values derived from or associated with `data`.
    """

    data: TData
    cache: TCache

WithIndices

Bases: Batched, Generic[I, TData]

Data paired with an :class:Index selecting a subset of elements.

Attributes:

Name Type Description
indices Index[I]

Index array mapping entries to labeled elements.

data TData

Associated data for the selected elements.

Source code in src/kups/core/data/wrappers.py
@dataclass
class WithIndices(Batched, Generic[I, TData]):
    """Data paired with an :class:`Index` selecting a subset of elements.

    Attributes:
        indices: Index array mapping entries to labeled elements.
        data: Associated data for the selected elements.
    """

    indices: Index[I]
    data: TData

    def map_data[NewData](
        self, f: Callable[[TData], NewData]
    ) -> WithIndices[I, NewData]:
        """Apply ``f`` to ``data``, keeping the same indices."""
        return WithIndices(self.indices, f(self.data))

map_data(f)

Apply f to data, keeping the same indices.

Source code in src/kups/core/data/wrappers.py
def map_data[NewData](
    self, f: Callable[[TData], NewData]
) -> WithIndices[I, NewData]:
    """Apply ``f`` to ``data``, keeping the same indices."""
    return WithIndices(self.indices, f(self.data))

subselect(target_ids, segment_ids, *, output_buffer_size, num_segments, is_sorted=False)

Find indices for elements belonging to target segments.

This function efficiently finds all occurrences of elements from specified target segment IDs within a segmented array, returning gather and scatter indices for data manipulation operations.

Parameters:

Name Type Description Default
target_ids Array

Array of segment IDs to search for.

required
segment_ids Array

Array mapping each element to its segment ID.

required
output_buffer_size Capacity[int]

Capacity object controlling output buffer size.

required
num_segments int

Total number of segments in the data.

required
is_sorted bool

Whether segment_ids is already sorted by segment.

False

Returns:

Type Description
SubselectResult

SubselectResult containing scatter_idxs and gather_idxs for indexing operations.

Source code in src/kups/core/utils/subselect.py
@functools.partial(jit, static_argnames=("num_segments", "is_sorted"))
def subselect(
    target_ids: Array,
    segment_ids: Array,
    *,
    output_buffer_size: Capacity[int],
    num_segments: int,
    is_sorted: bool = False,
) -> SubselectResult:
    """Find indices for elements belonging to target segments.

    This function efficiently finds all occurrences of elements from specified
    target segment IDs within a segmented array, returning gather and scatter
    indices for data manipulation operations.

    Args:
        target_ids: Array of segment IDs to search for.
        segment_ids: Array mapping each element to its segment ID.
        output_buffer_size: Capacity object controlling output buffer size.
        num_segments: Total number of segments in the data.
        is_sorted: Whether segment_ids is already sorted by segment.

    Returns:
        SubselectResult containing scatter_idxs and gather_idxs for indexing operations.
    """
    num_occurrences = jnp.bincount(segment_ids, length=num_segments)
    target_num_occ = num_occurrences.at[target_ids].get(mode="fill", fill_value=0)
    total_occurrences = jnp.sum(target_num_occ)
    output_buffer_size = output_buffer_size.generate_assertion(total_occurrences)
    # Compute gather indices for extracting matching elements
    gather_idxs = jnp.arange(output_buffer_size.size) + jnp.repeat(
        offsets_from_counts(num_occurrences)[target_ids]
        - offsets_from_counts(target_num_occ),
        target_num_occ,
        total_repeat_length=output_buffer_size.size,
    )
    # JAX cannot deal with indexing of zero-length arrays
    if gather_idxs.size == 0:
        return SubselectResult(jnp.array([], dtype=int), jnp.array([], dtype=int))
    if not is_sorted:
        gather_idxs = jnp.argsort(segment_ids)[gather_idxs]
    # Compute scatter indices for organizing results
    scatter_idxs = jnp.repeat(
        jnp.arange(len(target_ids)),
        target_num_occ,
        total_repeat_length=output_buffer_size.size,
    )
    mask = jnp.arange(output_buffer_size.size) < total_occurrences
    scatter_idxs = jnp.where(mask, scatter_idxs, len(target_ids))
    gather_idxs = jnp.where(mask, gather_idxs, len(segment_ids))
    return SubselectResult(scatter_idxs, gather_idxs)