Metadata-Version: 2.4
Name: onlinerake
Version: 0.1.1
Summary: Streaming survey raking via SGD and MWU
Author-email: Gaurav SOod <gsood07@gmail.com>
License: MIT
Requires-Python: >=3.8
Description-Content-Type: text/markdown
Requires-Dist: numpy>=1.21
Requires-Dist: pandas>=1.3
Provides-Extra: dev
Requires-Dist: pytest; extra == "dev"
Requires-Dist: black; extra == "dev"

# `onlinerake`

**Streaming survey raking in pure Python**

Modern online surveys and passive data collection streams generate
responses one record at a time.  Classic weighting methods such as
iterative proportional fitting (IPF, or “raking”) and calibration
weighting are inherently *batch* procedures: they reprocess the entire
dataset whenever a new case arrives.  The `onlinerake` package
provides **incremental**, per‑observation updates to survey weights so
that weighted margins track known population totals in real time.

The package implements two complementary algorithms:

* **SGD raking** – an additive update that performs stochastic
  gradient descent on a squared–error loss over the margins.  It
  produces smooth weight trajectories and maintains high effective
  sample size (ESS).
* **MWU raking** – a multiplicative update inspired by the
  multiplicative‑weights update rule.  It corresponds to mirror
  descent under the Kullback–Leibler divergence and yields weight
  distributions reminiscent of classic IPF.  However, it can produce
  heavier tails when the learning rate is large.

Both methods share the same API: call `.partial_fit(obs)` for each
incoming observation and inspect properties such as `.margins`, `.loss`
and `.effective_sample_size` to monitor progress.

## Installation

Clone or download this repository and install in editable mode:

```bash
git clone <repo-url>
cd onlinerake
pip install -e .
```

No external dependencies are required beyond `numpy` and `pandas`.

## Usage

```python
from onlinerake import OnlineRakingSGD, OnlineRakingMWU, Targets

# define target population margins (proportion of the population with indicator = 1)
targets = Targets(age=0.5, gender=0.5, education=0.4, region=0.3)

# instantiate a raker
raker = OnlineRakingSGD(targets, learning_rate=5.0)

# stream demographic observations
for obs in stream_of_dicts:
    raker.partial_fit(obs)
    print(raker.margins)  # current weighted margins

print("final effective sample size", raker.effective_sample_size)
```

To use the multiplicative‑weights version, replace
`OnlineRakingSGD` with `OnlineRakingMWU` and adjust the
`learning_rate` (a typical default is `1.0`).  See the docstrings
for full parameter descriptions.

## Simulation results

To understand the behaviour of the two update rules we simulated
three typical non‑stationary bias patterns: a **linear drift** in
demographic composition, a **sudden shift** halfway through the stream,
and an **oscillation** around the target frame.  For each scenario we
generated 300 observations per seed and averaged results over five
random seeds.  SGD used a learning rate of 5.0 and MWU used a
learning rate of 1.0 with three update steps per observation.  The
table below summarises the mean improvement in absolute margin error
relative to the unweighted baseline (positive values indicate an
improvement), the final effective sample size (ESS) and the mean final
loss (squared‑error on margins).  Higher ESS and larger improvements
are better.

| Scenario | Method | Age Imp (%) | Gender Imp (%) | Education Imp (%) | Region Imp (%) | Overall Imp (%) | Final ESS | Final Loss |
|---------|--------|-------------|---------------|------------------|---------------|----------------|---------:|-----------:|
| linear | SGD | 82.8 | 78.6 | 76.8 | 67.5 | 77.0 | 251.8 | 0.00147 |
| linear | MWU | 57.2 | 53.6 | 46.9 | 34.6 | 48.8 | 240.9 | 0.00676 |
| sudden | SGD | 82.9 | 82.3 | 79.6 | 63.5 | 79.5 | 225.5 | 0.00102 |
| sudden | MWU | 52.6 | 51.2 | 46.3 | 26.3 | 47.3 | 175.9 | 0.01235 |
| oscillating | SGD | 69.7 | 78.5 | 65.6 | 72.0 | 72.2 | 278.7 | 0.00023 |
| oscillating | MWU | 49.6 | 57.3 | 48.3 | 50.1 | 52.0 | 276.0 | 0.00048 |

**Interpretation**

* In all scenarios the online rakers dramatically reduce the margin
  errors relative to the unweighted baseline.  For example, in the
  sudden‑shift scenario the SGD raker reduces the average age error
  from 0.20 to about 0.03 (a 83% improvement).
* The SGD update consistently yields *higher* improvements and lower
  final loss than the MWU update, albeit at the cost of choosing a
  more aggressive learning rate.
* The MWU update, while less accurate in these settings, maintains
  comparable effective sample sizes and might be preferable when
  multiplicative adjustments are desired (e.g., when starting from
  unequal base weights).

You can reproduce these results or design new experiments by running

```bash
python -m onlinerake.simulation
```

from the repository root.  See the source of
`onlinerake/simulation.py` for details.

## Contributing

Pull requests are welcome!  Feel free to open issues if you find bugs
or have suggestions for new features, such as support for multi‑level
controls or adaptive learning‑rate schedules.
