From 57786a76b21f0a603f8cacd24e5626273236593a Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Wed, 27 Nov 2024 02:42:04 +0000 Subject: [PATCH] Deployed bbd54bc with MkDocs version: 1.6.1 --- .nojekyll | 0 404.html | 721 + advanced/custom-operators/index.html | 755 + advanced/extending-agents/index.html | 755 + advanced/performance-tuning/index.html | 755 + api-reference/cli/index.html | 1354 ++ api-reference/docetl/index.html | 8049 ++++++++ api-reference/operations/index.html | 16517 ++++++++++++++++ api-reference/optimizers/index.html | 11123 +++++++++++ api-reference/python/index.html | 3260 +++ assets/_mkdocstrings.css | 143 + assets/docetl-favicon-color.png | Bin 0 -> 1188 bytes assets/headerdiagram.png | Bin 0 -> 573787 bytes assets/images/favicon.png | Bin 0 -> 1870 bytes assets/javascripts/bundle.83f73b43.min.js | 16 + assets/javascripts/bundle.83f73b43.min.js.map | 7 + assets/javascripts/glightbox.min.js | 1 + assets/javascripts/lunr/min/lunr.ar.min.js | 1 + assets/javascripts/lunr/min/lunr.da.min.js | 18 + assets/javascripts/lunr/min/lunr.de.min.js | 18 + assets/javascripts/lunr/min/lunr.du.min.js | 18 + assets/javascripts/lunr/min/lunr.el.min.js | 1 + assets/javascripts/lunr/min/lunr.es.min.js | 18 + assets/javascripts/lunr/min/lunr.fi.min.js | 18 + assets/javascripts/lunr/min/lunr.fr.min.js | 18 + assets/javascripts/lunr/min/lunr.he.min.js | 1 + assets/javascripts/lunr/min/lunr.hi.min.js | 1 + assets/javascripts/lunr/min/lunr.hu.min.js | 18 + assets/javascripts/lunr/min/lunr.hy.min.js | 1 + assets/javascripts/lunr/min/lunr.it.min.js | 18 + assets/javascripts/lunr/min/lunr.ja.min.js | 1 + assets/javascripts/lunr/min/lunr.jp.min.js | 1 + assets/javascripts/lunr/min/lunr.kn.min.js | 1 + assets/javascripts/lunr/min/lunr.ko.min.js | 1 + assets/javascripts/lunr/min/lunr.multi.min.js | 1 + assets/javascripts/lunr/min/lunr.nl.min.js | 18 + assets/javascripts/lunr/min/lunr.no.min.js | 18 + assets/javascripts/lunr/min/lunr.pt.min.js | 18 + assets/javascripts/lunr/min/lunr.ro.min.js | 18 + assets/javascripts/lunr/min/lunr.ru.min.js | 18 + assets/javascripts/lunr/min/lunr.sa.min.js | 1 + .../lunr/min/lunr.stemmer.support.min.js | 1 + assets/javascripts/lunr/min/lunr.sv.min.js | 18 + assets/javascripts/lunr/min/lunr.ta.min.js | 1 + assets/javascripts/lunr/min/lunr.te.min.js | 1 + assets/javascripts/lunr/min/lunr.th.min.js | 1 + assets/javascripts/lunr/min/lunr.tr.min.js | 18 + assets/javascripts/lunr/min/lunr.vi.min.js | 1 + assets/javascripts/lunr/min/lunr.zh.min.js | 1 + assets/javascripts/lunr/tinyseg.js | 206 + assets/javascripts/lunr/wordcut.js | 6708 +++++++ .../workers/search.6ce7567c.min.js | 42 + .../workers/search.6ce7567c.min.js.map | 7 + assets/medical_transcripts.json | 437 + assets/readmefig.png | Bin 0 -> 377510 bytes assets/stylesheets/glightbox.min.css | 1 + assets/stylesheets/main.0253249f.min.css | 1 + assets/stylesheets/main.0253249f.min.css.map | 1 + assets/stylesheets/palette.06af60db.min.css | 1 + .../stylesheets/palette.06af60db.min.css.map | 1 + assets/tutorial/all-outputs.png | Bin 0 -> 602654 bytes assets/tutorial/console.png | Bin 0 -> 514925 bytes assets/tutorial/dataset-view.png | Bin 0 -> 671721 bytes assets/tutorial/initial-quotes.png | Bin 0 -> 584147 bytes assets/tutorial/operation-details.png | Bin 0 -> 335569 bytes assets/tutorial/playground-screenshot.png | Bin 0 -> 202071 bytes assets/tutorial/v2-quotes.png | Bin 0 -> 581948 bytes best-practices/index.html | 1990 ++ community/index.html | 941 + community/roadmap/index.html | 983 + concepts/operators/index.html | 2035 ++ concepts/optimization/index.html | 1897 ++ concepts/pipelines/index.html | 1902 ++ concepts/schemas/index.html | 1981 ++ .../annotating-legal-documents/index.html | 752 + .../characterizing-troll-behavior/index.html | 752 + examples/custom-parsing/index.html | 3096 +++ examples/mining-product-reviews/index.html | 1642 ++ examples/ollama/index.html | 1205 ++ .../presidential-debate-themes/index.html | 1618 ++ examples/rate-limiting/index.html | 1021 + examples/split-gather/index.html | 1550 ++ execution/running-pipelines/index.html | 923 + index.html | 1707 ++ installation/index.html | 1833 ++ objects.inv | Bin 0 -> 1317 bytes operators/cluster/index.html | 1970 ++ operators/code/index.html | 1965 ++ operators/equijoin/index.html | 1924 ++ operators/filter/index.html | 1903 ++ operators/gather/index.html | 2477 +++ operators/link-resolve/index.html | 900 + operators/map/index.html | 2225 +++ operators/parallel-map/index.html | 1961 ++ operators/reduce/index.html | 2261 +++ operators/resolve/index.html | 2023 ++ operators/sample/index.html | 1985 ++ operators/split/index.html | 2218 +++ operators/unnest/index.html | 2105 ++ optimization/configuration/index.html | 1879 ++ optimization/example/index.html | 2190 ++ optimization/overview/index.html | 1896 ++ optimization/python-api/index.html | 1743 ++ playground/features/index.html | 905 + playground/index.html | 1004 + playground/tutorial/index.html | 1053 + python-api/index.html | 1835 ++ search/search_index.json | 1 + sitemap.xml | 3 + sitemap.xml.gz | Bin 0 -> 127 bytes stylesheets/extra.css | 96 + tutorial/index.html | 1982 ++ 112 files changed, 119480 insertions(+) create mode 100644 .nojekyll create mode 100644 404.html create mode 100644 advanced/custom-operators/index.html create mode 100644 advanced/extending-agents/index.html create mode 100644 advanced/performance-tuning/index.html create mode 100644 api-reference/cli/index.html create mode 100644 api-reference/docetl/index.html create mode 100644 api-reference/operations/index.html create mode 100644 api-reference/optimizers/index.html create mode 100644 api-reference/python/index.html create mode 100644 assets/_mkdocstrings.css create mode 100644 assets/docetl-favicon-color.png create mode 100644 assets/headerdiagram.png create mode 100644 assets/images/favicon.png create mode 100644 assets/javascripts/bundle.83f73b43.min.js create mode 100644 assets/javascripts/bundle.83f73b43.min.js.map create mode 100644 assets/javascripts/glightbox.min.js create mode 100644 assets/javascripts/lunr/min/lunr.ar.min.js create mode 100644 assets/javascripts/lunr/min/lunr.da.min.js create mode 100644 assets/javascripts/lunr/min/lunr.de.min.js create mode 100644 assets/javascripts/lunr/min/lunr.du.min.js create mode 100644 assets/javascripts/lunr/min/lunr.el.min.js create mode 100644 assets/javascripts/lunr/min/lunr.es.min.js create mode 100644 assets/javascripts/lunr/min/lunr.fi.min.js create mode 100644 assets/javascripts/lunr/min/lunr.fr.min.js create mode 100644 assets/javascripts/lunr/min/lunr.he.min.js create mode 100644 assets/javascripts/lunr/min/lunr.hi.min.js create mode 100644 assets/javascripts/lunr/min/lunr.hu.min.js create mode 100644 assets/javascripts/lunr/min/lunr.hy.min.js create mode 100644 assets/javascripts/lunr/min/lunr.it.min.js create mode 100644 assets/javascripts/lunr/min/lunr.ja.min.js create mode 100644 assets/javascripts/lunr/min/lunr.jp.min.js create mode 100644 assets/javascripts/lunr/min/lunr.kn.min.js create mode 100644 assets/javascripts/lunr/min/lunr.ko.min.js create mode 100644 assets/javascripts/lunr/min/lunr.multi.min.js create mode 100644 assets/javascripts/lunr/min/lunr.nl.min.js create mode 100644 assets/javascripts/lunr/min/lunr.no.min.js create mode 100644 assets/javascripts/lunr/min/lunr.pt.min.js create mode 100644 assets/javascripts/lunr/min/lunr.ro.min.js create mode 100644 assets/javascripts/lunr/min/lunr.ru.min.js create mode 100644 assets/javascripts/lunr/min/lunr.sa.min.js create mode 100644 assets/javascripts/lunr/min/lunr.stemmer.support.min.js create mode 100644 assets/javascripts/lunr/min/lunr.sv.min.js create mode 100644 assets/javascripts/lunr/min/lunr.ta.min.js create mode 100644 assets/javascripts/lunr/min/lunr.te.min.js create mode 100644 assets/javascripts/lunr/min/lunr.th.min.js create mode 100644 assets/javascripts/lunr/min/lunr.tr.min.js create mode 100644 assets/javascripts/lunr/min/lunr.vi.min.js create mode 100644 assets/javascripts/lunr/min/lunr.zh.min.js create mode 100644 assets/javascripts/lunr/tinyseg.js create mode 100644 assets/javascripts/lunr/wordcut.js create mode 100644 assets/javascripts/workers/search.6ce7567c.min.js create mode 100644 assets/javascripts/workers/search.6ce7567c.min.js.map create mode 100644 assets/medical_transcripts.json create mode 100644 assets/readmefig.png create mode 100644 assets/stylesheets/glightbox.min.css create mode 100644 assets/stylesheets/main.0253249f.min.css create mode 100644 assets/stylesheets/main.0253249f.min.css.map create mode 100644 assets/stylesheets/palette.06af60db.min.css create mode 100644 assets/stylesheets/palette.06af60db.min.css.map create mode 100644 assets/tutorial/all-outputs.png create mode 100644 assets/tutorial/console.png create mode 100644 assets/tutorial/dataset-view.png create mode 100644 assets/tutorial/initial-quotes.png create mode 100644 assets/tutorial/operation-details.png create mode 100644 assets/tutorial/playground-screenshot.png create mode 100644 assets/tutorial/v2-quotes.png create mode 100644 best-practices/index.html create mode 100644 community/index.html create mode 100644 community/roadmap/index.html create mode 100644 concepts/operators/index.html create mode 100644 concepts/optimization/index.html create mode 100644 concepts/pipelines/index.html create mode 100644 concepts/schemas/index.html create mode 100644 examples/annotating-legal-documents/index.html create mode 100644 examples/characterizing-troll-behavior/index.html create mode 100644 examples/custom-parsing/index.html create mode 100644 examples/mining-product-reviews/index.html create mode 100644 examples/ollama/index.html create mode 100644 examples/presidential-debate-themes/index.html create mode 100644 examples/rate-limiting/index.html create mode 100644 examples/split-gather/index.html create mode 100644 execution/running-pipelines/index.html create mode 100644 index.html create mode 100644 installation/index.html create mode 100644 objects.inv create mode 100644 operators/cluster/index.html create mode 100644 operators/code/index.html create mode 100644 operators/equijoin/index.html create mode 100644 operators/filter/index.html create mode 100644 operators/gather/index.html create mode 100644 operators/link-resolve/index.html create mode 100644 operators/map/index.html create mode 100644 operators/parallel-map/index.html create mode 100644 operators/reduce/index.html create mode 100644 operators/resolve/index.html create mode 100644 operators/sample/index.html create mode 100644 operators/split/index.html create mode 100644 operators/unnest/index.html create mode 100644 optimization/configuration/index.html create mode 100644 optimization/example/index.html create mode 100644 optimization/overview/index.html create mode 100644 optimization/python-api/index.html create mode 100644 playground/features/index.html create mode 100644 playground/index.html create mode 100644 playground/tutorial/index.html create mode 100644 python-api/index.html create mode 100644 search/search_index.json create mode 100644 sitemap.xml create mode 100644 sitemap.xml.gz create mode 100644 stylesheets/extra.css create mode 100644 tutorial/index.html diff --git a/.nojekyll b/.nojekyll new file mode 100644 index 00000000..e69de29b diff --git a/404.html b/404.html new file mode 100644 index 00000000..1d6b31c7 --- /dev/null +++ b/404.html @@ -0,0 +1,721 @@ + + + + + + + + + + + + + + + + + + + docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ +

404 - Not found

+ +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/advanced/custom-operators/index.html b/advanced/custom-operators/index.html new file mode 100644 index 00000000..d0b0451b --- /dev/null +++ b/advanced/custom-operators/index.html @@ -0,0 +1,755 @@ + + + + + + + + + + + + + + + + + + + Custom operators - docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + + + + +

Custom operators

+ + + + + + + + + + + + + + + + + + +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/advanced/extending-agents/index.html b/advanced/extending-agents/index.html new file mode 100644 index 00000000..1a67c084 --- /dev/null +++ b/advanced/extending-agents/index.html @@ -0,0 +1,755 @@ + + + + + + + + + + + + + + + + + + + Extending agents - docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + + + + +

Extending agents

+ + + + + + + + + + + + + + + + + + +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/advanced/performance-tuning/index.html b/advanced/performance-tuning/index.html new file mode 100644 index 00000000..b78de10c --- /dev/null +++ b/advanced/performance-tuning/index.html @@ -0,0 +1,755 @@ + + + + + + + + + + + + + + + + + + + Performance tuning - docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + + + + +

Performance tuning

+ + + + + + + + + + + + + + + + + + +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/api-reference/cli/index.html b/api-reference/cli/index.html new file mode 100644 index 00000000..8f10cfb4 --- /dev/null +++ b/api-reference/cli/index.html @@ -0,0 +1,1354 @@ + + + + + + + + + + + + + + + + + + + + + + + docetl.cli - docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + + + + +

docetl.cli

+ +
+ + +

+ docetl.cli.run(yaml_file=typer.Argument(..., help='Path to the YAML file containing the pipeline configuration'), max_threads=typer.Option(None, help='Maximum number of threads to use for running operations')) + +

+ + +
+ +

Run the configuration specified in the YAML file.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ yaml_file + + Path + +
+

Path to the YAML file containing the pipeline configuration.

+
+
+ Argument(..., help='Path to the YAML file containing the pipeline configuration') +
+ max_threads + + Optional[int] + +
+

Maximum number of threads to use for running operations.

+
+
+ Option(None, help='Maximum number of threads to use for running operations') +
+ +
+ Source code in docetl/cli.py +
55
+56
+57
+58
+59
+60
+61
+62
+63
+64
+65
+66
+67
+68
+69
+70
+71
+72
+73
+74
+75
+76
+77
+78
+79
+80
@app.command()
+def run(
+    yaml_file: Path = typer.Argument(
+        ..., help="Path to the YAML file containing the pipeline configuration"
+    ),
+    max_threads: Optional[int] = typer.Option(
+        None, help="Maximum number of threads to use for running operations"
+    ),
+):
+    """
+    Run the configuration specified in the YAML file.
+
+    Args:
+        yaml_file (Path): Path to the YAML file containing the pipeline configuration.
+        max_threads (Optional[int]): Maximum number of threads to use for running operations.
+    """
+    # Get the current working directory (where the user called the command)
+    cwd = os.getcwd()
+
+    # Load .env file from the current working directory
+    env_file = os.path.join(cwd, ".env")
+    if os.path.exists(env_file):
+        load_dotenv(env_file)
+
+    runner = DSLRunner.from_yaml(str(yaml_file), max_threads=max_threads)
+    runner.load_run_save()
+
+
+
+ +
+ +
+ + +

+ docetl.cli.build(yaml_file=typer.Argument(..., help='Path to the YAML file containing the pipeline configuration'), max_threads=typer.Option(None, help='Maximum number of threads to use for running operations'), model=typer.Option('gpt-4o', help='Model to use for optimization'), resume=typer.Option(False, help='Resume optimization from a previous build that may have failed'), timeout=typer.Option(60, help='Timeout for optimization operations in seconds')) + +

+ + +
+ +

Build and optimize the configuration specified in the YAML file.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ yaml_file + + Path + +
+

Path to the YAML file containing the pipeline configuration.

+
+
+ Argument(..., help='Path to the YAML file containing the pipeline configuration') +
+ max_threads + + Optional[int] + +
+

Maximum number of threads to use for running operations.

+
+
+ Option(None, help='Maximum number of threads to use for running operations') +
+ model + + str + +
+

Model to use for optimization. Defaults to "gpt-4o".

+
+
+ Option('gpt-4o', help='Model to use for optimization') +
+ resume + + bool + +
+

Whether to resume optimization from a previous run. Defaults to False.

+
+
+ Option(False, help='Resume optimization from a previous build that may have failed') +
+ timeout + + int + +
+

Timeout for optimization operations in seconds. Defaults to 60.

+
+
+ Option(60, help='Timeout for optimization operations in seconds') +
+ +
+ Source code in docetl/cli.py +
15
+16
+17
+18
+19
+20
+21
+22
+23
+24
+25
+26
+27
+28
+29
+30
+31
+32
+33
+34
+35
+36
+37
+38
+39
+40
+41
+42
+43
+44
+45
+46
+47
+48
+49
+50
+51
+52
@app.command()
+def build(
+    yaml_file: Path = typer.Argument(
+        ..., help="Path to the YAML file containing the pipeline configuration"
+    ),
+    max_threads: Optional[int] = typer.Option(
+        None, help="Maximum number of threads to use for running operations"
+    ),
+    model: str = typer.Option("gpt-4o", help="Model to use for optimization"),
+    resume: bool = typer.Option(
+        False, help="Resume optimization from a previous build that may have failed"
+    ),
+    timeout: int = typer.Option(
+        60, help="Timeout for optimization operations in seconds"
+    ),
+):
+    """
+    Build and optimize the configuration specified in the YAML file.
+
+    Args:
+        yaml_file (Path): Path to the YAML file containing the pipeline configuration.
+        max_threads (Optional[int]): Maximum number of threads to use for running operations.
+        model (str): Model to use for optimization. Defaults to "gpt-4o".
+        resume (bool): Whether to resume optimization from a previous run. Defaults to False.
+        timeout (int): Timeout for optimization operations in seconds. Defaults to 60.
+    """
+    # Get the current working directory (where the user called the command)
+    cwd = os.getcwd()
+
+    # Load .env file from the current working directory
+    env_file = os.path.join(cwd, ".env")
+    if os.path.exists(env_file):
+        load_dotenv(env_file)
+
+    runner = DSLRunner.from_yaml(str(yaml_file), max_threads=max_threads)
+    runner.optimize(
+        save=True, return_pipeline=False, model=model, resume=resume, timeout=timeout
+    )
+
+
+
+ +
+ +
+ + +

+ docetl.cli.clear_cache() + +

+ + +
+ +

Clear the LLM cache stored on disk.

+ +
+ Source code in docetl/cli.py +
83
+84
+85
+86
+87
+88
@app.command()
+def clear_cache():
+    """
+    Clear the LLM cache stored on disk.
+    """
+    cc()
+
+
+
+ +
+ + + + + + + + + + + + + +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/api-reference/docetl/index.html b/api-reference/docetl/index.html new file mode 100644 index 00000000..2c6a54cb --- /dev/null +++ b/api-reference/docetl/index.html @@ -0,0 +1,8049 @@ + + + + + + + + + + + + + + + + + + + + + + + docetl - docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + + + + + +
+
+ + + + + + + +

docetl

+ +
+ + + +

+ docetl.DSLRunner + + +

+ + +
+

+ Bases: ConfigWrapper

+ + +

This class is responsible for running DocETL pipelines. It manages datasets, executes pipeline steps, and tracks +the cost of operations.

+ + +

Attributes:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescription
config + Dict + +
+

The loaded configuration from the YAML file.

+
+
default_model + str + +
+

The default language model to use for operations.

+
+
max_threads + int + +
+

Maximum number of threads for parallel processing.

+
+
console + Console + +
+

Rich console for output formatting.

+
+
datasets + Dict + +
+

Storage for loaded datasets.

+
+
+ + + + + + +
+ Source code in docetl/runner.py +
 27
+ 28
+ 29
+ 30
+ 31
+ 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
+470
+471
+472
+473
+474
+475
+476
+477
+478
+479
+480
+481
+482
+483
+484
+485
+486
+487
+488
+489
+490
+491
+492
+493
+494
+495
+496
+497
+498
+499
+500
+501
+502
+503
+504
+505
+506
+507
+508
+509
+510
+511
+512
+513
+514
class DSLRunner(ConfigWrapper):
+    """
+    This class is responsible for running DocETL pipelines. It manages datasets, executes pipeline steps, and tracks
+    the cost of operations.
+
+    Attributes:
+        config (Dict): The loaded configuration from the YAML file.
+        default_model (str): The default language model to use for operations.
+        max_threads (int): Maximum number of threads for parallel processing.
+        console (Console): Rich console for output formatting.
+        datasets (Dict): Storage for loaded datasets.
+    """
+
+    @classproperty
+    def schema(cls):
+        # Accessing the schema loads all operations, so only do this
+        # when we actually need it...
+        # Yes, this means DSLRunner.schema isn't really accessible to
+        # static type checkers. But it /is/ available for dynamic
+        # checking, and for generating json schema.
+
+        OpType = functools.reduce(
+            lambda a, b: a | b, [op.schema for op in get_operations().values()]
+        )
+        # More pythonic implementation of the above, but only works in python 3.11:
+        # OpType = Union[*[op.schema for op in get_operations().values()]]
+
+        class Pipeline(BaseModel):
+            config: Optional[dict[str, Any]]
+            parsing_tools: Optional[list[schemas.ParsingTool]]
+            datasets: Dict[str, schemas.Dataset]
+            operations: list[OpType]
+            pipeline: schemas.PipelineSpec
+
+        return Pipeline
+
+    @classproperty
+    def json_schema(cls):
+        return cls.schema.model_json_schema()
+
+    def __init__(self, config: Dict, max_threads: int = None, **kwargs):
+        """
+        Initialize the DSLRunner with a YAML configuration file.
+
+        Args:
+            max_threads (int, optional): Maximum number of threads to use. Defaults to None.
+        """
+        super().__init__(
+            config,
+            base_name=kwargs.pop("base_name", None),
+            yaml_file_suffix=kwargs.pop("yaml_file_suffix", None),
+            max_threads=max_threads,
+            **kwargs,
+        )
+        self.datasets = {}
+
+        self.intermediate_dir = (
+            self.config.get("pipeline", {}).get("output", {}).get("intermediate_dir")
+        )
+
+        # Create parsing tool map
+        self.parsing_tool_map = create_parsing_tool_map(
+            self.config.get("parsing_tools", None)
+        )
+
+        self.syntax_check()
+
+        op_map = {op["name"]: op for op in self.config["operations"]}
+
+        # Hash each pipeline step/operation
+        # for each step op, hash the code of each op up until and (including that op)
+        self.step_op_hashes = defaultdict(dict)
+        for step in self.config["pipeline"]["steps"]:
+            for idx, op in enumerate(step["operations"]):
+                op_name = op if isinstance(op, str) else list(op.keys())[0]
+
+                all_ops_until_and_including_current = [
+                    op_map[prev_op] for prev_op in step["operations"][:idx]
+                ] + [op_map[op_name]] + [self.config.get("system_prompt", {})]
+                # If there's no model in the op, add the default model
+                for op in all_ops_until_and_including_current:
+                    if "model" not in op:
+                        op["model"] = self.default_model
+
+                all_ops_str = json.dumps(all_ops_until_and_including_current)
+                self.step_op_hashes[step["name"]][op_name] = hashlib.sha256(
+                    all_ops_str.encode()
+                ).hexdigest()
+
+    def get_output_path(self, require=False):
+        output_path = self.config.get("pipeline", {}).get("output", {}).get("path")
+        if output_path:
+            if not (
+                output_path.lower().endswith(".json")
+                or output_path.lower().endswith(".csv")
+            ):
+                raise ValueError(
+                    f"Output path '{output_path}' is not a JSON or CSV file. Please provide a path ending with '.json' or '.csv'."
+                )
+        elif require:
+            raise ValueError(
+                "No output path specified in the configuration. Please provide an output path ending with '.json' or '.csv' in the configuration to use the save() method."
+            )
+
+        return output_path
+
+    def syntax_check(self):
+        """
+        Perform a syntax check on all operations defined in the configuration.
+
+        This method validates each operation by attempting to instantiate it.
+        If any operation fails to instantiate, a ValueError is raised.
+
+        Raises:
+            ValueError: If any operation fails the syntax check.
+        """
+        self.console.rule("[yellow]Syntax Check[/yellow]")
+        self.console.print(
+            "[yellow]Performing syntax check on all operations...[/yellow]"
+        )
+
+        # Just validate that it's a json file if specified
+        self.get_output_path()
+
+        for operation_config in self.config["operations"]:
+            operation = operation_config["name"]
+            operation_type = operation_config["type"]
+
+            try:
+                operation_class = get_operation(operation_type)
+                operation_class(
+                    self,
+                    operation_config,
+                    self.default_model,
+                    self.max_threads,
+                    self.console,
+                )
+            except Exception as e:
+                raise ValueError(
+                    f"Syntax check failed for operation '{operation}': {str(e)}"
+                )
+
+        self.console.print("[green]Syntax check passed for all operations.[/green]")
+
+    def find_operation(self, op_name: str) -> Dict:
+        for operation_config in self.config["operations"]:
+            if operation_config["name"] == op_name:
+                return operation_config
+        raise ValueError(f"Operation '{op_name}' not found in configuration.")
+
+    def load_run_save(self) -> float:
+        """
+        Execute the entire pipeline defined in the configuration.
+
+        This method loads datasets, executes each step in the pipeline, saves the output,
+        and returns the total cost of execution.
+
+        Returns:
+            float: The total cost of executing the pipeline.
+        """
+
+        # Fail early if we can't save the output...
+        self.get_output_path(require=True)
+
+        self.console.rule("[bold blue]Pipeline Execution[/bold blue]")
+        start_time = time.time()
+
+        output, total_cost = self.run(self.load())
+        self.save(output)
+
+        self.console.rule("[bold green]Execution Summary[/bold green]")
+        self.console.print(f"[bold green]Total cost: [green]${total_cost:.2f}[/green]")
+        self.console.print(
+            f"[bold green]Total time: [green]{time.time() - start_time:.2f} seconds[/green]"
+        )
+
+        return total_cost
+
+    def run(self, datasets) -> float:
+        """
+        Execute the entire pipeline defined in the configuration on some data.
+
+        Args:
+           datasets (dict[str, Dataset | List[Dict]]): input datasets to transform
+
+        Returns:
+            (List[Dict], float): The transformed data and the total cost of execution.
+        """
+        self.datasets = {
+            name: (
+                dataset
+                if isinstance(dataset, Dataset)
+                else Dataset(self, "memory", dataset)
+            )
+            for name, dataset in datasets.items()
+        }
+        total_cost = 0
+        for step in self.config["pipeline"]["steps"]:
+            step_name = step["name"]
+            input_data = (
+                self.datasets[step["input"]].load() if "input" in step else None
+            )
+            output_data, step_cost = self.execute_step(step, input_data)
+            self.datasets[step_name] = Dataset(self, "memory", output_data)
+            flush_cache(self.console)
+            total_cost += step_cost
+            self.console.log(
+                f"Step [cyan]{step_name}[/cyan] completed. Cost: [green]${step_cost:.2f}[/green]"
+            )
+
+        # Save the self.step_op_hashes to a file if self.intermediate_dir exists
+        if self.intermediate_dir:
+            with open(
+                os.path.join(self.intermediate_dir, ".docetl_intermediate_config.json"),
+                "w",
+            ) as f:
+                json.dump(self.step_op_hashes, f)
+
+        return (
+            self.datasets[self.config["pipeline"]["steps"][-1]["name"]].load(),
+            total_cost,
+        )
+
+    def load(self):
+        """
+        Load all datasets defined in the configuration.
+
+        This method creates Dataset objects for each dataset in the configuration.
+
+        Raises:
+            ValueError: If an unsupported dataset type is encountered.
+        """
+        self.console.rule("[cyan]Loading Datasets[/cyan]")
+        datasets = {}
+        for name, dataset_config in self.config["datasets"].items():
+            if dataset_config["type"] == "file":
+                datasets[name] = Dataset(
+                    self,
+                    "file",
+                    dataset_config["path"],
+                    source="local",
+                    parsing=dataset_config.get("parsing", []),
+                    user_defined_parsing_tool_map=self.parsing_tool_map,
+                )
+                self.console.print(f"Loaded dataset: [bold]{name}[/bold]")
+            else:
+                raise ValueError(f"Unsupported dataset type: {dataset_config['type']}")
+        return datasets
+
+    def save(self, data: List[Dict]):
+        """
+        Save the final output of the pipeline.
+
+        Args:
+            data (List[Dict]): The data to be saved.
+
+        Raises:
+            ValueError: If an unsupported output type is specified in the configuration.
+        """
+        self.get_output_path(require=True)
+
+        self.console.rule("[cyan]Saving Output[/cyan]")
+        output_config = self.config["pipeline"]["output"]
+        if output_config["type"] == "file":
+            # Create the directory if it doesn't exist
+            if os.path.dirname(output_config["path"]):
+                os.makedirs(os.path.dirname(output_config["path"]), exist_ok=True)
+            if output_config["path"].lower().endswith(".json"):
+                with open(output_config["path"], "w") as file:
+                    json.dump(data, file, indent=2)
+            else:  # CSV
+                import csv
+
+                with open(output_config["path"], "w", newline="") as file:
+                    writer = csv.DictWriter(file, fieldnames=data[0].keys())
+                    writer.writeheader()
+                    writer.writerows(data)
+            self.console.print(
+                f"[green italic]💾 Output saved to {output_config['path']}[/green italic]"
+            )
+        else:
+            raise ValueError(f"Unsupported output type: {output_config['type']}")
+
+    def execute_step(
+        self, step: Dict, input_data: Optional[List[Dict]]
+    ) -> Tuple[List[Dict], float]:
+        """
+        Execute a single step in the pipeline.
+
+        This method runs all operations defined for a step, updating the progress
+        and calculating the cost.
+
+        Args:
+            step (Dict): The step configuration.
+            input_data (Optional[List[Dict]]): Input data for the step.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the output data and the total cost of the step.
+        """
+        self.console.rule(f"[bold blue]Executing Step: {step['name']}[/bold blue]")
+        total_cost = 0
+        for operation in step["operations"]:
+            if isinstance(operation, dict):
+                operation_name = list(operation.keys())[0]
+                operation_config = self.find_operation(operation_name)
+            else:
+                operation_name = operation
+                operation_config = {}
+
+            # Load from checkpoint if it exists
+            attempted_input_data = self._load_from_checkpoint_if_exists(
+                step["name"], operation_name
+            )
+            if attempted_input_data is not None:
+                input_data = attempted_input_data
+                self.console.print(
+                    f"[green]✓ [italic]Loaded saved data for operation '{operation_name}' in step '{step['name']}'[/italic][/green]"
+                )
+                continue
+
+            op_object = self.find_operation(operation_name).copy()
+            op_object.update(operation_config)
+
+            # If sample is set, sample the input data
+            if op_object.get("sample"):
+                if input_data is None:
+                    input_data = self.datasets[step["input"]].sample(op_object["sample"], False)
+                else:
+                    input_data = input_data[: op_object["sample"]]
+
+            with self.console.status("[bold]Running Operation:[/bold]") as status:
+                status.update(f"Type: [cyan]{op_object['type']}[/cyan]")
+                status.update(f"Name: [cyan]{op_object.get('name', 'Unnamed')}[/cyan]")
+                self.status = status
+
+                operation_class = get_operation(op_object["type"])
+                operation_instance = operation_class(
+                    self,
+                    op_object,
+                    self.default_model,
+                    self.max_threads,
+                    self.console,
+                    self.status,
+                )
+                if op_object["type"] == "equijoin":
+                    left_data = self.datasets[op_object["left"]].load()
+                    right_data = self.datasets[op_object["right"]].load()
+                    input_data, cost = operation_instance.execute(left_data, right_data)
+                else:
+                    input_data, cost = operation_instance.execute(input_data)
+                total_cost += cost
+                self.console.log(
+                    f"\tOperation [cyan]{operation_name}[/cyan] completed. Cost: [green]${cost:.2f}[/green]"
+                )
+
+            # Checkpoint after each operation
+            if self.intermediate_dir:
+                self._save_checkpoint(step["name"], operation_name, input_data)
+
+                # Load existing step op hash, if exists, merge self.step_op_hashes[step["name"]][operation_name] into it
+                # Save the step op hash
+                intermediate_config_path = os.path.join(
+                    self.intermediate_dir, ".docetl_intermediate_config.json"
+                )
+                if os.path.exists(intermediate_config_path):
+                    with open(intermediate_config_path, "r") as f:
+                        existing_config = json.load(f)
+                else:
+                    existing_config = {}
+
+                if step["name"] not in existing_config:
+                    existing_config[step["name"]] = {}
+                existing_config[step["name"]][operation_name] = self.step_op_hashes[
+                    step["name"]
+                ][operation_name]
+
+                # Resave
+                with open(intermediate_config_path, "w") as f:
+                    json.dump(existing_config, f, indent=2)
+
+        return input_data, total_cost
+
+    def _load_from_checkpoint_if_exists(
+        self, step_name: str, operation_name: str
+    ) -> Optional[List[Dict]]:
+        if self.intermediate_dir is None:
+            return None
+
+        intermediate_config_path = os.path.join(
+            self.intermediate_dir, ".docetl_intermediate_config.json"
+        )
+
+        if not os.path.exists(intermediate_config_path):
+            return None
+
+        # See if the checkpoint config is the same as the current step op hash
+        with open(intermediate_config_path, "r") as f:
+            intermediate_config = json.load(f)
+
+        if (
+            intermediate_config.get(step_name, {}).get(operation_name, "")
+            != self.step_op_hashes[step_name][operation_name]
+        ):
+            return None
+
+        checkpoint_path = os.path.join(
+            self.intermediate_dir, step_name, f"{operation_name}.json"
+        )
+        # check if checkpoint exists
+        if os.path.exists(checkpoint_path):
+            if f"{step_name}_{operation_name}" not in self.datasets:
+                self.datasets[f"{step_name}_{operation_name}"] = Dataset(
+                    self, "file", checkpoint_path, "local"
+                )
+            return self.datasets[f"{step_name}_{operation_name}"].load()
+        return None
+
+    def clear_intermediate(self) -> None:
+        """
+        Clear the intermediate directory.
+        """
+        # Remove the intermediate directory
+        if self.intermediate_dir:
+            shutil.rmtree(self.intermediate_dir)
+            return
+
+        raise ValueError("Intermediate directory not set. Cannot clear intermediate.")
+
+    def _save_checkpoint(self, step_name: str, operation_name: str, data: List[Dict]):
+        """
+        Save a checkpoint of the current data after an operation.
+
+        This method creates a JSON file containing the current state of the data
+        after an operation has been executed. The checkpoint is saved in a directory
+        structure that reflects the step and operation names.
+
+        Args:
+            step_name (str): The name of the current step in the pipeline.
+            operation_name (str): The name of the operation that was just executed.
+            data (List[Dict]): The current state of the data to be checkpointed.
+
+        Note:
+            The checkpoint is saved only if a checkpoint directory has been specified
+            when initializing the DSLRunner.
+        """
+        checkpoint_path = os.path.join(
+            self.intermediate_dir, step_name, f"{operation_name}.json"
+        )
+        if os.path.dirname(checkpoint_path):
+            os.makedirs(os.path.dirname(checkpoint_path), exist_ok=True)
+        with open(checkpoint_path, "w") as f:
+            json.dump(data, f)
+
+        self.console.print(
+            f"[green]✓ [italic]Intermediate saved for operation '{operation_name}' in step '{step_name}' at {checkpoint_path}[/italic][/green]"
+        )
+
+    def should_optimize(self, step_name: str, op_name: str, **kwargs) -> Tuple[str, float, List[Dict[str, Any]], List[Dict[str, Any]]]:
+        builder = Optimizer(self, **kwargs)
+        return builder.should_optimize(step_name, op_name)
+
+    def optimize(
+        self,
+        save: bool = False,
+        return_pipeline: bool = True,
+        **kwargs,
+    ) -> Tuple[Union[Dict, "DSLRunner"], float]:
+
+        builder = Optimizer(
+            self,
+            **kwargs,
+        )
+        cost = builder.optimize()
+
+        # Dump via json
+        # import json
+        # with open(f"{self.base_name}_optimizer_output.json", "wb") as f:
+        #     json.dump(builder.captured_output.optimizer_output, f)
+
+
+        if save:
+            builder.save_optimized_config(f"{self.base_name}_opt.yaml")
+            self.optimized_config_path = f"{self.base_name}_opt.yaml"
+
+        if return_pipeline:
+            return DSLRunner(builder.clean_optimized_config(), self.max_threads), cost
+
+        return builder.clean_optimized_config(), cost
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ __init__(config, max_threads=None, **kwargs) + +

+ + +
+ +

Initialize the DSLRunner with a YAML configuration file.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ max_threads + + int + +
+

Maximum number of threads to use. Defaults to None.

+
+
+ None +
+ +
+ Source code in docetl/runner.py +
 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
def __init__(self, config: Dict, max_threads: int = None, **kwargs):
+    """
+    Initialize the DSLRunner with a YAML configuration file.
+
+    Args:
+        max_threads (int, optional): Maximum number of threads to use. Defaults to None.
+    """
+    super().__init__(
+        config,
+        base_name=kwargs.pop("base_name", None),
+        yaml_file_suffix=kwargs.pop("yaml_file_suffix", None),
+        max_threads=max_threads,
+        **kwargs,
+    )
+    self.datasets = {}
+
+    self.intermediate_dir = (
+        self.config.get("pipeline", {}).get("output", {}).get("intermediate_dir")
+    )
+
+    # Create parsing tool map
+    self.parsing_tool_map = create_parsing_tool_map(
+        self.config.get("parsing_tools", None)
+    )
+
+    self.syntax_check()
+
+    op_map = {op["name"]: op for op in self.config["operations"]}
+
+    # Hash each pipeline step/operation
+    # for each step op, hash the code of each op up until and (including that op)
+    self.step_op_hashes = defaultdict(dict)
+    for step in self.config["pipeline"]["steps"]:
+        for idx, op in enumerate(step["operations"]):
+            op_name = op if isinstance(op, str) else list(op.keys())[0]
+
+            all_ops_until_and_including_current = [
+                op_map[prev_op] for prev_op in step["operations"][:idx]
+            ] + [op_map[op_name]] + [self.config.get("system_prompt", {})]
+            # If there's no model in the op, add the default model
+            for op in all_ops_until_and_including_current:
+                if "model" not in op:
+                    op["model"] = self.default_model
+
+            all_ops_str = json.dumps(all_ops_until_and_including_current)
+            self.step_op_hashes[step["name"]][op_name] = hashlib.sha256(
+                all_ops_str.encode()
+            ).hexdigest()
+
+
+
+ +
+ +
+ + +

+ clear_intermediate() + +

+ + +
+ +

Clear the intermediate directory.

+ +
+ Source code in docetl/runner.py +
444
+445
+446
+447
+448
+449
+450
+451
+452
+453
def clear_intermediate(self) -> None:
+    """
+    Clear the intermediate directory.
+    """
+    # Remove the intermediate directory
+    if self.intermediate_dir:
+        shutil.rmtree(self.intermediate_dir)
+        return
+
+    raise ValueError("Intermediate directory not set. Cannot clear intermediate.")
+
+
+
+ +
+ +
+ + +

+ execute_step(step, input_data) + +

+ + +
+ +

Execute a single step in the pipeline.

+

This method runs all operations defined for a step, updating the progress +and calculating the cost.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ step + + Dict + +
+

The step configuration.

+
+
+ required +
+ input_data + + Optional[List[Dict]] + +
+

Input data for the step.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[List[Dict], float] + +
+

Tuple[List[Dict], float]: A tuple containing the output data and the total cost of the step.

+
+
+ +
+ Source code in docetl/runner.py +
310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
def execute_step(
+    self, step: Dict, input_data: Optional[List[Dict]]
+) -> Tuple[List[Dict], float]:
+    """
+    Execute a single step in the pipeline.
+
+    This method runs all operations defined for a step, updating the progress
+    and calculating the cost.
+
+    Args:
+        step (Dict): The step configuration.
+        input_data (Optional[List[Dict]]): Input data for the step.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the output data and the total cost of the step.
+    """
+    self.console.rule(f"[bold blue]Executing Step: {step['name']}[/bold blue]")
+    total_cost = 0
+    for operation in step["operations"]:
+        if isinstance(operation, dict):
+            operation_name = list(operation.keys())[0]
+            operation_config = self.find_operation(operation_name)
+        else:
+            operation_name = operation
+            operation_config = {}
+
+        # Load from checkpoint if it exists
+        attempted_input_data = self._load_from_checkpoint_if_exists(
+            step["name"], operation_name
+        )
+        if attempted_input_data is not None:
+            input_data = attempted_input_data
+            self.console.print(
+                f"[green]✓ [italic]Loaded saved data for operation '{operation_name}' in step '{step['name']}'[/italic][/green]"
+            )
+            continue
+
+        op_object = self.find_operation(operation_name).copy()
+        op_object.update(operation_config)
+
+        # If sample is set, sample the input data
+        if op_object.get("sample"):
+            if input_data is None:
+                input_data = self.datasets[step["input"]].sample(op_object["sample"], False)
+            else:
+                input_data = input_data[: op_object["sample"]]
+
+        with self.console.status("[bold]Running Operation:[/bold]") as status:
+            status.update(f"Type: [cyan]{op_object['type']}[/cyan]")
+            status.update(f"Name: [cyan]{op_object.get('name', 'Unnamed')}[/cyan]")
+            self.status = status
+
+            operation_class = get_operation(op_object["type"])
+            operation_instance = operation_class(
+                self,
+                op_object,
+                self.default_model,
+                self.max_threads,
+                self.console,
+                self.status,
+            )
+            if op_object["type"] == "equijoin":
+                left_data = self.datasets[op_object["left"]].load()
+                right_data = self.datasets[op_object["right"]].load()
+                input_data, cost = operation_instance.execute(left_data, right_data)
+            else:
+                input_data, cost = operation_instance.execute(input_data)
+            total_cost += cost
+            self.console.log(
+                f"\tOperation [cyan]{operation_name}[/cyan] completed. Cost: [green]${cost:.2f}[/green]"
+            )
+
+        # Checkpoint after each operation
+        if self.intermediate_dir:
+            self._save_checkpoint(step["name"], operation_name, input_data)
+
+            # Load existing step op hash, if exists, merge self.step_op_hashes[step["name"]][operation_name] into it
+            # Save the step op hash
+            intermediate_config_path = os.path.join(
+                self.intermediate_dir, ".docetl_intermediate_config.json"
+            )
+            if os.path.exists(intermediate_config_path):
+                with open(intermediate_config_path, "r") as f:
+                    existing_config = json.load(f)
+            else:
+                existing_config = {}
+
+            if step["name"] not in existing_config:
+                existing_config[step["name"]] = {}
+            existing_config[step["name"]][operation_name] = self.step_op_hashes[
+                step["name"]
+            ][operation_name]
+
+            # Resave
+            with open(intermediate_config_path, "w") as f:
+                json.dump(existing_config, f, indent=2)
+
+    return input_data, total_cost
+
+
+
+ +
+ +
+ + +

+ load() + +

+ + +
+ +

Load all datasets defined in the configuration.

+

This method creates Dataset objects for each dataset in the configuration.

+ + +

Raises:

+ + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If an unsupported dataset type is encountered.

+
+
+ +
+ Source code in docetl/runner.py +
250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
def load(self):
+    """
+    Load all datasets defined in the configuration.
+
+    This method creates Dataset objects for each dataset in the configuration.
+
+    Raises:
+        ValueError: If an unsupported dataset type is encountered.
+    """
+    self.console.rule("[cyan]Loading Datasets[/cyan]")
+    datasets = {}
+    for name, dataset_config in self.config["datasets"].items():
+        if dataset_config["type"] == "file":
+            datasets[name] = Dataset(
+                self,
+                "file",
+                dataset_config["path"],
+                source="local",
+                parsing=dataset_config.get("parsing", []),
+                user_defined_parsing_tool_map=self.parsing_tool_map,
+            )
+            self.console.print(f"Loaded dataset: [bold]{name}[/bold]")
+        else:
+            raise ValueError(f"Unsupported dataset type: {dataset_config['type']}")
+    return datasets
+
+
+
+ +
+ +
+ + +

+ load_run_save() + +

+ + +
+ +

Execute the entire pipeline defined in the configuration.

+

This method loads datasets, executes each step in the pipeline, saves the output, +and returns the total cost of execution.

+ + +

Returns:

+ + + + + + + + + + + + + +
Name TypeDescription
float + float + +
+

The total cost of executing the pipeline.

+
+
+ +
+ Source code in docetl/runner.py +
177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
def load_run_save(self) -> float:
+    """
+    Execute the entire pipeline defined in the configuration.
+
+    This method loads datasets, executes each step in the pipeline, saves the output,
+    and returns the total cost of execution.
+
+    Returns:
+        float: The total cost of executing the pipeline.
+    """
+
+    # Fail early if we can't save the output...
+    self.get_output_path(require=True)
+
+    self.console.rule("[bold blue]Pipeline Execution[/bold blue]")
+    start_time = time.time()
+
+    output, total_cost = self.run(self.load())
+    self.save(output)
+
+    self.console.rule("[bold green]Execution Summary[/bold green]")
+    self.console.print(f"[bold green]Total cost: [green]${total_cost:.2f}[/green]")
+    self.console.print(
+        f"[bold green]Total time: [green]{time.time() - start_time:.2f} seconds[/green]"
+    )
+
+    return total_cost
+
+
+
+ +
+ +
+ + +

+ run(datasets) + +

+ + +
+ +

Execute the entire pipeline defined in the configuration on some data.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ datasets + + dict[str, Dataset | List[Dict]] + +
+

input datasets to transform

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ (List[Dict], float) + +
+

The transformed data and the total cost of execution.

+
+
+ +
+ Source code in docetl/runner.py +
205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
def run(self, datasets) -> float:
+    """
+    Execute the entire pipeline defined in the configuration on some data.
+
+    Args:
+       datasets (dict[str, Dataset | List[Dict]]): input datasets to transform
+
+    Returns:
+        (List[Dict], float): The transformed data and the total cost of execution.
+    """
+    self.datasets = {
+        name: (
+            dataset
+            if isinstance(dataset, Dataset)
+            else Dataset(self, "memory", dataset)
+        )
+        for name, dataset in datasets.items()
+    }
+    total_cost = 0
+    for step in self.config["pipeline"]["steps"]:
+        step_name = step["name"]
+        input_data = (
+            self.datasets[step["input"]].load() if "input" in step else None
+        )
+        output_data, step_cost = self.execute_step(step, input_data)
+        self.datasets[step_name] = Dataset(self, "memory", output_data)
+        flush_cache(self.console)
+        total_cost += step_cost
+        self.console.log(
+            f"Step [cyan]{step_name}[/cyan] completed. Cost: [green]${step_cost:.2f}[/green]"
+        )
+
+    # Save the self.step_op_hashes to a file if self.intermediate_dir exists
+    if self.intermediate_dir:
+        with open(
+            os.path.join(self.intermediate_dir, ".docetl_intermediate_config.json"),
+            "w",
+        ) as f:
+            json.dump(self.step_op_hashes, f)
+
+    return (
+        self.datasets[self.config["pipeline"]["steps"][-1]["name"]].load(),
+        total_cost,
+    )
+
+
+
+ +
+ +
+ + +

+ save(data) + +

+ + +
+ +

Save the final output of the pipeline.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ data + + List[Dict] + +
+

The data to be saved.

+
+
+ required +
+ + +

Raises:

+ + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If an unsupported output type is specified in the configuration.

+
+
+ +
+ Source code in docetl/runner.py +
276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
def save(self, data: List[Dict]):
+    """
+    Save the final output of the pipeline.
+
+    Args:
+        data (List[Dict]): The data to be saved.
+
+    Raises:
+        ValueError: If an unsupported output type is specified in the configuration.
+    """
+    self.get_output_path(require=True)
+
+    self.console.rule("[cyan]Saving Output[/cyan]")
+    output_config = self.config["pipeline"]["output"]
+    if output_config["type"] == "file":
+        # Create the directory if it doesn't exist
+        if os.path.dirname(output_config["path"]):
+            os.makedirs(os.path.dirname(output_config["path"]), exist_ok=True)
+        if output_config["path"].lower().endswith(".json"):
+            with open(output_config["path"], "w") as file:
+                json.dump(data, file, indent=2)
+        else:  # CSV
+            import csv
+
+            with open(output_config["path"], "w", newline="") as file:
+                writer = csv.DictWriter(file, fieldnames=data[0].keys())
+                writer.writeheader()
+                writer.writerows(data)
+        self.console.print(
+            f"[green italic]💾 Output saved to {output_config['path']}[/green italic]"
+        )
+    else:
+        raise ValueError(f"Unsupported output type: {output_config['type']}")
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Perform a syntax check on all operations defined in the configuration.

+

This method validates each operation by attempting to instantiate it. +If any operation fails to instantiate, a ValueError is raised.

+ + +

Raises:

+ + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If any operation fails the syntax check.

+
+
+ +
+ Source code in docetl/runner.py +
133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
def syntax_check(self):
+    """
+    Perform a syntax check on all operations defined in the configuration.
+
+    This method validates each operation by attempting to instantiate it.
+    If any operation fails to instantiate, a ValueError is raised.
+
+    Raises:
+        ValueError: If any operation fails the syntax check.
+    """
+    self.console.rule("[yellow]Syntax Check[/yellow]")
+    self.console.print(
+        "[yellow]Performing syntax check on all operations...[/yellow]"
+    )
+
+    # Just validate that it's a json file if specified
+    self.get_output_path()
+
+    for operation_config in self.config["operations"]:
+        operation = operation_config["name"]
+        operation_type = operation_config["type"]
+
+        try:
+            operation_class = get_operation(operation_type)
+            operation_class(
+                self,
+                operation_config,
+                self.default_model,
+                self.max_threads,
+                self.console,
+            )
+        except Exception as e:
+            raise ValueError(
+                f"Syntax check failed for operation '{operation}': {str(e)}"
+            )
+
+    self.console.print("[green]Syntax check passed for all operations.[/green]")
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.Optimizer + + +

+ + +
+ + + + + + + +
+ Source code in docetl/builder.py +
  81
+  82
+  83
+  84
+  85
+  86
+  87
+  88
+  89
+  90
+  91
+  92
+  93
+  94
+  95
+  96
+  97
+  98
+  99
+ 100
+ 101
+ 102
+ 103
+ 104
+ 105
+ 106
+ 107
+ 108
+ 109
+ 110
+ 111
+ 112
+ 113
+ 114
+ 115
+ 116
+ 117
+ 118
+ 119
+ 120
+ 121
+ 122
+ 123
+ 124
+ 125
+ 126
+ 127
+ 128
+ 129
+ 130
+ 131
+ 132
+ 133
+ 134
+ 135
+ 136
+ 137
+ 138
+ 139
+ 140
+ 141
+ 142
+ 143
+ 144
+ 145
+ 146
+ 147
+ 148
+ 149
+ 150
+ 151
+ 152
+ 153
+ 154
+ 155
+ 156
+ 157
+ 158
+ 159
+ 160
+ 161
+ 162
+ 163
+ 164
+ 165
+ 166
+ 167
+ 168
+ 169
+ 170
+ 171
+ 172
+ 173
+ 174
+ 175
+ 176
+ 177
+ 178
+ 179
+ 180
+ 181
+ 182
+ 183
+ 184
+ 185
+ 186
+ 187
+ 188
+ 189
+ 190
+ 191
+ 192
+ 193
+ 194
+ 195
+ 196
+ 197
+ 198
+ 199
+ 200
+ 201
+ 202
+ 203
+ 204
+ 205
+ 206
+ 207
+ 208
+ 209
+ 210
+ 211
+ 212
+ 213
+ 214
+ 215
+ 216
+ 217
+ 218
+ 219
+ 220
+ 221
+ 222
+ 223
+ 224
+ 225
+ 226
+ 227
+ 228
+ 229
+ 230
+ 231
+ 232
+ 233
+ 234
+ 235
+ 236
+ 237
+ 238
+ 239
+ 240
+ 241
+ 242
+ 243
+ 244
+ 245
+ 246
+ 247
+ 248
+ 249
+ 250
+ 251
+ 252
+ 253
+ 254
+ 255
+ 256
+ 257
+ 258
+ 259
+ 260
+ 261
+ 262
+ 263
+ 264
+ 265
+ 266
+ 267
+ 268
+ 269
+ 270
+ 271
+ 272
+ 273
+ 274
+ 275
+ 276
+ 277
+ 278
+ 279
+ 280
+ 281
+ 282
+ 283
+ 284
+ 285
+ 286
+ 287
+ 288
+ 289
+ 290
+ 291
+ 292
+ 293
+ 294
+ 295
+ 296
+ 297
+ 298
+ 299
+ 300
+ 301
+ 302
+ 303
+ 304
+ 305
+ 306
+ 307
+ 308
+ 309
+ 310
+ 311
+ 312
+ 313
+ 314
+ 315
+ 316
+ 317
+ 318
+ 319
+ 320
+ 321
+ 322
+ 323
+ 324
+ 325
+ 326
+ 327
+ 328
+ 329
+ 330
+ 331
+ 332
+ 333
+ 334
+ 335
+ 336
+ 337
+ 338
+ 339
+ 340
+ 341
+ 342
+ 343
+ 344
+ 345
+ 346
+ 347
+ 348
+ 349
+ 350
+ 351
+ 352
+ 353
+ 354
+ 355
+ 356
+ 357
+ 358
+ 359
+ 360
+ 361
+ 362
+ 363
+ 364
+ 365
+ 366
+ 367
+ 368
+ 369
+ 370
+ 371
+ 372
+ 373
+ 374
+ 375
+ 376
+ 377
+ 378
+ 379
+ 380
+ 381
+ 382
+ 383
+ 384
+ 385
+ 386
+ 387
+ 388
+ 389
+ 390
+ 391
+ 392
+ 393
+ 394
+ 395
+ 396
+ 397
+ 398
+ 399
+ 400
+ 401
+ 402
+ 403
+ 404
+ 405
+ 406
+ 407
+ 408
+ 409
+ 410
+ 411
+ 412
+ 413
+ 414
+ 415
+ 416
+ 417
+ 418
+ 419
+ 420
+ 421
+ 422
+ 423
+ 424
+ 425
+ 426
+ 427
+ 428
+ 429
+ 430
+ 431
+ 432
+ 433
+ 434
+ 435
+ 436
+ 437
+ 438
+ 439
+ 440
+ 441
+ 442
+ 443
+ 444
+ 445
+ 446
+ 447
+ 448
+ 449
+ 450
+ 451
+ 452
+ 453
+ 454
+ 455
+ 456
+ 457
+ 458
+ 459
+ 460
+ 461
+ 462
+ 463
+ 464
+ 465
+ 466
+ 467
+ 468
+ 469
+ 470
+ 471
+ 472
+ 473
+ 474
+ 475
+ 476
+ 477
+ 478
+ 479
+ 480
+ 481
+ 482
+ 483
+ 484
+ 485
+ 486
+ 487
+ 488
+ 489
+ 490
+ 491
+ 492
+ 493
+ 494
+ 495
+ 496
+ 497
+ 498
+ 499
+ 500
+ 501
+ 502
+ 503
+ 504
+ 505
+ 506
+ 507
+ 508
+ 509
+ 510
+ 511
+ 512
+ 513
+ 514
+ 515
+ 516
+ 517
+ 518
+ 519
+ 520
+ 521
+ 522
+ 523
+ 524
+ 525
+ 526
+ 527
+ 528
+ 529
+ 530
+ 531
+ 532
+ 533
+ 534
+ 535
+ 536
+ 537
+ 538
+ 539
+ 540
+ 541
+ 542
+ 543
+ 544
+ 545
+ 546
+ 547
+ 548
+ 549
+ 550
+ 551
+ 552
+ 553
+ 554
+ 555
+ 556
+ 557
+ 558
+ 559
+ 560
+ 561
+ 562
+ 563
+ 564
+ 565
+ 566
+ 567
+ 568
+ 569
+ 570
+ 571
+ 572
+ 573
+ 574
+ 575
+ 576
+ 577
+ 578
+ 579
+ 580
+ 581
+ 582
+ 583
+ 584
+ 585
+ 586
+ 587
+ 588
+ 589
+ 590
+ 591
+ 592
+ 593
+ 594
+ 595
+ 596
+ 597
+ 598
+ 599
+ 600
+ 601
+ 602
+ 603
+ 604
+ 605
+ 606
+ 607
+ 608
+ 609
+ 610
+ 611
+ 612
+ 613
+ 614
+ 615
+ 616
+ 617
+ 618
+ 619
+ 620
+ 621
+ 622
+ 623
+ 624
+ 625
+ 626
+ 627
+ 628
+ 629
+ 630
+ 631
+ 632
+ 633
+ 634
+ 635
+ 636
+ 637
+ 638
+ 639
+ 640
+ 641
+ 642
+ 643
+ 644
+ 645
+ 646
+ 647
+ 648
+ 649
+ 650
+ 651
+ 652
+ 653
+ 654
+ 655
+ 656
+ 657
+ 658
+ 659
+ 660
+ 661
+ 662
+ 663
+ 664
+ 665
+ 666
+ 667
+ 668
+ 669
+ 670
+ 671
+ 672
+ 673
+ 674
+ 675
+ 676
+ 677
+ 678
+ 679
+ 680
+ 681
+ 682
+ 683
+ 684
+ 685
+ 686
+ 687
+ 688
+ 689
+ 690
+ 691
+ 692
+ 693
+ 694
+ 695
+ 696
+ 697
+ 698
+ 699
+ 700
+ 701
+ 702
+ 703
+ 704
+ 705
+ 706
+ 707
+ 708
+ 709
+ 710
+ 711
+ 712
+ 713
+ 714
+ 715
+ 716
+ 717
+ 718
+ 719
+ 720
+ 721
+ 722
+ 723
+ 724
+ 725
+ 726
+ 727
+ 728
+ 729
+ 730
+ 731
+ 732
+ 733
+ 734
+ 735
+ 736
+ 737
+ 738
+ 739
+ 740
+ 741
+ 742
+ 743
+ 744
+ 745
+ 746
+ 747
+ 748
+ 749
+ 750
+ 751
+ 752
+ 753
+ 754
+ 755
+ 756
+ 757
+ 758
+ 759
+ 760
+ 761
+ 762
+ 763
+ 764
+ 765
+ 766
+ 767
+ 768
+ 769
+ 770
+ 771
+ 772
+ 773
+ 774
+ 775
+ 776
+ 777
+ 778
+ 779
+ 780
+ 781
+ 782
+ 783
+ 784
+ 785
+ 786
+ 787
+ 788
+ 789
+ 790
+ 791
+ 792
+ 793
+ 794
+ 795
+ 796
+ 797
+ 798
+ 799
+ 800
+ 801
+ 802
+ 803
+ 804
+ 805
+ 806
+ 807
+ 808
+ 809
+ 810
+ 811
+ 812
+ 813
+ 814
+ 815
+ 816
+ 817
+ 818
+ 819
+ 820
+ 821
+ 822
+ 823
+ 824
+ 825
+ 826
+ 827
+ 828
+ 829
+ 830
+ 831
+ 832
+ 833
+ 834
+ 835
+ 836
+ 837
+ 838
+ 839
+ 840
+ 841
+ 842
+ 843
+ 844
+ 845
+ 846
+ 847
+ 848
+ 849
+ 850
+ 851
+ 852
+ 853
+ 854
+ 855
+ 856
+ 857
+ 858
+ 859
+ 860
+ 861
+ 862
+ 863
+ 864
+ 865
+ 866
+ 867
+ 868
+ 869
+ 870
+ 871
+ 872
+ 873
+ 874
+ 875
+ 876
+ 877
+ 878
+ 879
+ 880
+ 881
+ 882
+ 883
+ 884
+ 885
+ 886
+ 887
+ 888
+ 889
+ 890
+ 891
+ 892
+ 893
+ 894
+ 895
+ 896
+ 897
+ 898
+ 899
+ 900
+ 901
+ 902
+ 903
+ 904
+ 905
+ 906
+ 907
+ 908
+ 909
+ 910
+ 911
+ 912
+ 913
+ 914
+ 915
+ 916
+ 917
+ 918
+ 919
+ 920
+ 921
+ 922
+ 923
+ 924
+ 925
+ 926
+ 927
+ 928
+ 929
+ 930
+ 931
+ 932
+ 933
+ 934
+ 935
+ 936
+ 937
+ 938
+ 939
+ 940
+ 941
+ 942
+ 943
+ 944
+ 945
+ 946
+ 947
+ 948
+ 949
+ 950
+ 951
+ 952
+ 953
+ 954
+ 955
+ 956
+ 957
+ 958
+ 959
+ 960
+ 961
+ 962
+ 963
+ 964
+ 965
+ 966
+ 967
+ 968
+ 969
+ 970
+ 971
+ 972
+ 973
+ 974
+ 975
+ 976
+ 977
+ 978
+ 979
+ 980
+ 981
+ 982
+ 983
+ 984
+ 985
+ 986
+ 987
+ 988
+ 989
+ 990
+ 991
+ 992
+ 993
+ 994
+ 995
+ 996
+ 997
+ 998
+ 999
+1000
+1001
+1002
+1003
+1004
+1005
+1006
+1007
+1008
+1009
+1010
+1011
+1012
+1013
+1014
+1015
+1016
+1017
+1018
+1019
+1020
+1021
+1022
+1023
+1024
+1025
+1026
+1027
+1028
+1029
+1030
+1031
+1032
+1033
+1034
+1035
+1036
+1037
+1038
+1039
+1040
+1041
+1042
+1043
+1044
+1045
+1046
+1047
+1048
+1049
+1050
+1051
+1052
+1053
+1054
+1055
+1056
+1057
+1058
+1059
+1060
+1061
+1062
+1063
+1064
+1065
+1066
+1067
+1068
+1069
+1070
+1071
+1072
+1073
+1074
+1075
+1076
+1077
+1078
+1079
+1080
+1081
+1082
+1083
+1084
+1085
+1086
+1087
+1088
+1089
+1090
+1091
+1092
+1093
+1094
+1095
+1096
+1097
+1098
+1099
+1100
+1101
+1102
+1103
+1104
+1105
+1106
+1107
+1108
+1109
+1110
+1111
+1112
+1113
+1114
+1115
+1116
+1117
+1118
+1119
+1120
+1121
+1122
+1123
+1124
+1125
+1126
+1127
+1128
+1129
+1130
+1131
+1132
+1133
+1134
+1135
+1136
+1137
+1138
+1139
+1140
+1141
+1142
+1143
+1144
+1145
+1146
+1147
+1148
+1149
+1150
+1151
+1152
+1153
+1154
+1155
+1156
+1157
+1158
+1159
+1160
+1161
+1162
+1163
+1164
+1165
+1166
+1167
+1168
+1169
+1170
+1171
+1172
+1173
+1174
+1175
+1176
+1177
+1178
+1179
+1180
+1181
+1182
+1183
+1184
+1185
+1186
+1187
+1188
+1189
+1190
+1191
+1192
+1193
+1194
+1195
+1196
+1197
+1198
+1199
+1200
+1201
+1202
+1203
+1204
+1205
+1206
+1207
+1208
+1209
+1210
+1211
+1212
+1213
+1214
+1215
+1216
+1217
+1218
+1219
+1220
+1221
+1222
+1223
+1224
+1225
+1226
+1227
+1228
+1229
+1230
+1231
+1232
+1233
+1234
+1235
+1236
+1237
+1238
+1239
+1240
+1241
+1242
+1243
+1244
+1245
+1246
+1247
+1248
+1249
+1250
+1251
+1252
+1253
+1254
+1255
+1256
+1257
+1258
+1259
+1260
+1261
+1262
+1263
+1264
+1265
+1266
+1267
+1268
+1269
+1270
+1271
+1272
+1273
+1274
+1275
+1276
+1277
+1278
+1279
+1280
+1281
+1282
+1283
+1284
+1285
+1286
+1287
+1288
+1289
+1290
+1291
+1292
+1293
+1294
+1295
+1296
+1297
+1298
+1299
+1300
+1301
+1302
+1303
+1304
+1305
+1306
+1307
+1308
+1309
+1310
+1311
+1312
+1313
+1314
+1315
+1316
+1317
+1318
+1319
+1320
+1321
+1322
+1323
+1324
+1325
+1326
+1327
+1328
+1329
+1330
+1331
+1332
+1333
+1334
+1335
+1336
+1337
+1338
+1339
+1340
+1341
+1342
+1343
+1344
+1345
+1346
+1347
+1348
+1349
+1350
+1351
+1352
+1353
+1354
+1355
+1356
+1357
+1358
+1359
+1360
+1361
+1362
+1363
+1364
+1365
+1366
+1367
+1368
+1369
+1370
+1371
+1372
+1373
+1374
+1375
+1376
+1377
+1378
+1379
+1380
+1381
+1382
+1383
+1384
+1385
+1386
+1387
+1388
+1389
+1390
+1391
+1392
+1393
+1394
+1395
+1396
+1397
+1398
+1399
+1400
+1401
+1402
+1403
+1404
+1405
+1406
+1407
+1408
+1409
+1410
+1411
+1412
+1413
+1414
+1415
+1416
+1417
+1418
+1419
+1420
+1421
+1422
+1423
+1424
+1425
+1426
+1427
+1428
+1429
+1430
+1431
+1432
+1433
+1434
+1435
+1436
+1437
+1438
+1439
+1440
+1441
+1442
+1443
+1444
+1445
+1446
+1447
+1448
+1449
+1450
+1451
+1452
+1453
+1454
+1455
+1456
+1457
+1458
+1459
+1460
+1461
+1462
+1463
+1464
+1465
+1466
+1467
+1468
+1469
+1470
+1471
+1472
+1473
+1474
+1475
+1476
+1477
+1478
+1479
+1480
+1481
+1482
+1483
+1484
+1485
+1486
+1487
+1488
+1489
+1490
+1491
+1492
+1493
+1494
+1495
+1496
+1497
+1498
+1499
+1500
+1501
+1502
+1503
+1504
+1505
+1506
+1507
+1508
+1509
+1510
+1511
+1512
+1513
+1514
+1515
+1516
+1517
+1518
+1519
class Optimizer:
+
+    def __init__(
+        self,
+        runner: "DSLRunner",
+        model: str = "gpt-4o",
+        resume: bool = False,
+        timeout: int = 60,
+    ):
+        """
+        Initialize the Optimizer class.
+
+        This method sets up the optimizer with the given configuration file and parameters.
+        It loads the configuration, initializes the console for output, sets up the LLM client,
+        and prepares various attributes for optimization.
+
+        Args:
+            yaml_file (str): Path to the YAML configuration file.
+            max_threads (Optional[int]): Maximum number of threads to use for parallel processing.
+                If None, it will be set to (number of CPUs * 4).
+            model (str): The name of the language model to use. Defaults to "gpt-4o".
+            resume (bool): Whether to resume optimization from a previous run. Defaults to False.
+            timeout (int): Timeout in seconds for operations. Defaults to 60.
+
+        Attributes:
+            yaml_file_path (str): Stores the path to the YAML file.
+            config (Dict): Stores the loaded configuration from the YAML file.
+            console (Console): Rich console for formatted output.
+            optimized_config (Dict): A copy of the original config to be optimized.
+            llm_client (LLMClient): Client for interacting with the language model.
+            max_threads (int): Maximum number of threads for parallel processing.
+            operations_cost (float): Tracks the total cost of operations.
+            timeout (int): Timeout for operations in seconds.
+            selectivities (defaultdict): Stores selectivity information for operations.
+                Selectivity is the ratio of output size to input size for an operation.
+                It's used to estimate how much data will flow through the pipeline after
+                each operation, which helps in optimizing subsequent operations and
+                determining appropriate sample sizes. For example, a selectivity of 0.5
+                means an operation halves the size of its input data.
+            datasets (Dict): Stores loaded datasets.
+
+        The method also calls print_optimizer_config() to display the initial configuration.
+        """
+        self.config = runner.config
+        self.console = runner.console
+        self.max_threads = runner.max_threads
+
+        self.base_name = runner.base_name
+        self.yaml_file_suffix = runner.yaml_file_suffix
+        self.config = runner.config
+        self.runner = runner
+        self.status = runner.status
+
+        self.optimized_config = copy.deepcopy(self.config)
+        self.llm_client = LLMClient(model)
+        self.operations_cost = 0
+        self.timeout = timeout
+        self.selectivities = defaultdict(dict)
+        self.samples_taken = defaultdict(dict)
+        self.resume = resume
+        self.captured_output = CapturedOutput()
+
+        home_dir = os.environ.get("DOCETL_HOME_DIR", os.path.expanduser("~"))
+        cache_dir = os.path.join(home_dir, f".docetl/cache/{runner.yaml_file_suffix}")
+        os.makedirs(cache_dir, exist_ok=True)
+        self.datasets = DatasetOnDisk(dir=cache_dir, console=self.console)
+        self.optimized_ops_path = f"{cache_dir}/optimized_ops"
+
+        # Update sample size map
+        self.sample_size_map = SAMPLE_SIZE_MAP
+        if self.config.get("optimizer_config", {}).get("sample_sizes", {}):
+            self.sample_size_map.update(self.config["optimizer_config"]["sample_sizes"])
+
+        self.step_op_to_optimized_ops = {}
+
+        self.print_optimizer_config()
+
+    def find_operation(self, op_name: str, config: Optional[Dict] = None) -> Dict:
+        if not config:
+            config = self.config
+        for operation_config in config["operations"]:
+            if operation_config["name"] == op_name:
+                return operation_config
+        raise ValueError(f"Operation '{op_name}' not found in configuration.")
+
+    def syntax_check(self):
+        """
+        Perform a syntax check on all operations defined in the configuration.
+
+        This method validates each operation by attempting to instantiate it.
+        If any operation fails to instantiate, a ValueError is raised.
+
+        Raises:
+            ValueError: If any operation fails the syntax check.
+        """
+        for operation_config in self.config["operations"]:
+            operation = operation_config["name"]
+            operation_type = operation_config["type"]
+
+            try:
+                operation_class = get_operation(operation_type)
+                operation_class(
+                    self.runner,
+                    operation_config,
+                    self.config.get("default_model", "gpt-4o-mini"),
+                    self.max_threads,
+                    console=self.console,
+                )
+            except Exception as e:
+                raise ValueError(
+                    f"Syntax check failed for operation '{operation}': {str(e)}"
+                )
+
+        self.console.log("[green]Syntax check passed for all operations.[/green]")
+
+    def print_optimizer_config(self):
+        """
+        Print the current configuration of the optimizer.
+
+        This method uses the Rich console to display a formatted output of the optimizer's
+        configuration. It includes details such as the YAML file path, sample sizes for
+        different operation types, maximum number of threads, the language model being used,
+        and the timeout setting.
+
+        The output is color-coded and formatted for easy readability, with a header and
+        separator lines to clearly delineate the configuration information.
+        """
+        self.console.rule("[bold cyan]Optimizer Configuration[/bold cyan]")
+        self.console.log(f"[yellow]Sample Size:[/yellow] {self.sample_size_map}")
+        self.console.log(f"[yellow]Max Threads:[/yellow] {self.max_threads}")
+        self.console.log(f"[yellow]Model:[/yellow] {self.llm_client.model}")
+        self.console.log(f"[yellow]Timeout:[/yellow] {self.timeout} seconds")
+
+    def compute_sample_size(
+        self,
+        step_name: str,
+        step_ops: List[str],
+        op_config: Dict[str, Any],
+    ) -> int:
+        """
+        Compute the sample size necessary for optimizing given operation based on upstream operations.
+
+        This method calculates an appropriate sample size for an operation, taking into
+        account the selectivities of upstream operations in the same step. It uses a
+        predefined sample size map (SAMPLE_SIZE_MAP) as a starting point.
+
+        For example, if we have a 'map' operation with a default sample size of 10,
+        and one upstream operation with a selectivity of 0.5, the computed sample size for the upstream operation would be:
+        10 / 0.5 = 20
+
+        This ensures that after applying the selectivity of the upstream operation,
+        we still have a representative sample size for the current operation.
+
+        Args:
+            step_name (str): The name of the current step in the pipeline.
+            step_ops (List[str]): A list of all operations in the current step.
+            op_config (Dict[str, Any]): The configuration dictionary for the current operation.
+
+        Returns:
+            int: The computed sample size for the operation.
+
+        The method works as follows:
+        1. If there are no upstream operations, it returns the default sample size for the operation type.
+        2. Otherwise, it starts with the default sample size and adjusts it based on the selectivities
+           of upstream operations.
+        3. It iterates through upstream operations in reverse order, dividing the sample size by
+           each operation's selectivity.
+        4. The final result is rounded to the nearest integer.
+
+        Raises:
+            ValueError: If the selectivity for any upstream operation is not found.
+
+        Note:
+            - The method assumes that selectivities for all upstream operations have been
+              previously computed and stored in self.selectivities.
+            - The sample size is always at least 1, even after all adjustments.
+        """
+        # If an equijoin, load the default. Equijoins are always first
+        if op_config.get("type") == "equijoin":
+            return SAMPLE_SIZE_MAP.get(op_config.get("type"))
+
+        # If there are no upstream operations, use the default sample_size
+        upstream_ops = []
+        for step_op in step_ops:
+            if step_op != op_config.get("name"):
+                if step_op in self.step_op_to_optimized_ops:
+                    upstream_ops.extend(self.step_op_to_optimized_ops[step_op])
+                else:
+                    upstream_ops.append(step_op)
+            else:
+                break
+
+        if len(upstream_ops) == 0:
+            return self.sample_size_map.get(op_config.get("type"), float("inf"))
+
+        # Otherwise, compute the sample size based on the upstream operations
+        sample_size = self.sample_size_map.get(op_config.get("type"), 100)
+        for op in reversed(upstream_ops):
+            # Use the selectivity of the upstream operation to compute the sample size
+            if op not in self.selectivities[step_name]:
+                raise ValueError(
+                    f"Selectivity for operation {op} not found in selectivities. Other ops are {self.selectivities[step_name]}"
+                )
+
+            sample_size = sample_size / self.selectivities[step_name].get(op)
+
+        return int(math.ceil(sample_size))
+
+    def _insert_empty_resolve_operations(self):
+        """
+        Determines whether to insert resolve operations in the pipeline.
+
+        This method iterates through each step in the pipeline and checks if there's a reduce
+        operation that follows a map operation with no resolver in between. If such a case is
+        found, it synthesizes an empty resolver operation and inserts it into the pipeline.
+
+        The method modifies the pipeline configuration in-place.
+
+        Returns:
+            None
+
+        Side effects:
+        - Modifies self.config["pipeline"]["steps"] by potentially inserting new resolve operations.
+        - Adds new resolve operations to self.config["operations"] if necessary.
+        """
+        for i, step in enumerate(self.config["pipeline"]["steps"]):
+            operations = step.get("operations", [])
+            has_map = False
+            has_reduce = False
+            has_resolve = False
+            map_op = None
+            reduce_op = None
+
+            for op in operations:
+                if isinstance(op, dict):
+                    op = list(op.keys())[0]
+                op_config = self.find_operation(op)
+                op_type = op_config["type"]
+                if op_type == "map":
+                    has_map = True
+                    map_op = op
+                elif op_type == "reduce" and op_config.get("synthesize_resolve", True):
+                    reduce_key = op_config.get("reduce_key", "_all")
+                    if isinstance(reduce_key, str):
+                        reduce_key = [reduce_key]
+                    if "_all" not in reduce_key:
+                        has_reduce = True
+                        reduce_op = op
+                elif op_type == "resolve":
+                    has_resolve = True
+
+            if has_map and has_reduce and not has_resolve:
+                # Synthesize an empty resolver
+                self.console.log(
+                    "[yellow]Synthesizing empty resolver operation:[/yellow]"
+                )
+                self.console.log(
+                    f"  • [cyan]Reduce operation:[/cyan] [bold]{reduce_op}[/bold]"
+                )
+                self.console.log(f"  • [cyan]Step:[/cyan] [bold]{step['name']}[/bold]")
+
+                new_resolve_op = f"synthesized_resolve_{i}"
+                reduce_key = self.find_operation(reduce_op).get("reduce_key")
+                if isinstance(reduce_key, str):
+                    reduce_key = [reduce_key]
+                self.config["operations"].append(
+                    {
+                        "name": new_resolve_op,
+                        "type": "resolve",
+                        "empty": True,
+                        "optimize": True,
+                        "embedding_model": "text-embedding-3-small",
+                        "resolution_model": self.config.get(
+                            "default_model", "gpt-4o-mini"
+                        ),
+                        "comparison_model": self.config.get(
+                            "default_model", "gpt-4o-mini"
+                        ),
+                        "_intermediates": {
+                            "map_prompt": self.find_operation(map_op).get("prompt"),
+                            "reduce_key": reduce_key,
+                        },
+                    }
+                )
+
+                # Insert the new resolve operation before the reduce operation
+                reduce_index = next(
+                    i
+                    for i, op in enumerate(operations)
+                    if self.find_operation(op).get("type") == "reduce"
+                )
+                operations.insert(reduce_index, new_resolve_op)
+
+                has_resolve = True
+
+            self.config["pipeline"]["steps"][i]["operations"] = operations
+
+        # Update the pipeline configuration
+        self.config["pipeline"]["steps"] = self.config["pipeline"]["steps"]
+
+    def _add_map_prompts_to_reduce_operations(self):
+        """
+        Add relevant map prompts to reduce operations based on their reduce keys.
+
+        This method iterates through all map operations to create a dictionary mapping
+        output schema keys to map prompts. It then loops through reduce operations,
+        adding the relevant map prompts based on the reduce keys and output schema.
+
+        Side effects:
+        - Modifies reduce operations in self.config["operations"] by adding map prompts.
+        """
+        # Create a dictionary mapping output schema keys to map prompts
+        output_key_to_prompt = {}
+        for op_config in self.config["operations"]:
+            if op_config.get("type") == "map":
+                output_schema = op_config.get("output", {}).get("schema", {})
+                prompt = op_config.get("prompt", "")
+                for key in output_schema.keys():
+                    output_key_to_prompt[key] = prompt
+
+        # Add relevant map prompts to reduce operations
+        for op_config in self.config["operations"]:
+            if op_config.get("type") == "reduce":
+                reduce_keys = op_config.get("reduce_key", [])
+                if isinstance(reduce_keys, str):
+                    reduce_keys = [reduce_keys]
+
+                relevant_prompts = []
+                for key in reduce_keys:
+                    if key in output_key_to_prompt:
+                        relevant_prompts.append(output_key_to_prompt[key])
+
+                if relevant_prompts:
+                    op_config["_intermediates"] = op_config.get("_intermediates", {})
+                    op_config["_intermediates"]["last_map_prompt"] = relevant_prompts[
+                        -1
+                    ]
+
+    def _load_optimized_ops(self):
+        """
+        Load the optimized operations from disk.
+        """
+        if os.path.exists(self.optimized_ops_path):
+            for filename in os.listdir(self.optimized_ops_path):
+                if filename.endswith(".json"):
+                    original_op_name = filename[:-5]  # Remove '.json' from the filename
+                    with open(
+                        os.path.join(self.optimized_ops_path, filename), "r"
+                    ) as f:
+                        optimized_ops = json.load(f)
+
+                    # Update the config with the optimized operations
+                    if original_op_name in [
+                        op["name"] for op in self.config["operations"]
+                    ]:
+                        # Update the config with the optimized operations
+                        # First, remove all operations that are already in the config with the same name
+                        self.config["operations"] = [
+                            op
+                            for op in self.config["operations"]
+                            if op["name"] != original_op_name
+                        ]
+
+                        for op in optimized_ops:
+                            op["optimize"] = False
+                            self.config["operations"].append(op)
+
+                        # Update the step operations
+                        for step in self.config["pipeline"]["steps"]:
+                            if original_op_name in step["operations"]:
+                                index = step["operations"].index(original_op_name)
+                                step["operations"] = (
+                                    step["operations"][:index]
+                                    + [op["name"] for op in optimized_ops]
+                                    + step["operations"][index + 1 :]
+                                )
+
+                    self.console.log(
+                        f"Loaded optimized operations for {original_op_name}"
+                    )
+
+            self.console.log("[green]Finished loading optimized operations[/green]")
+
+            # Print out the operations for each step
+            self.console.log("[bold blue]Operations for each step:[/bold blue]")
+            for step in self.config["pipeline"]["steps"]:
+                step_name = step.get("name")
+                operations = step.get("operations", [])
+                self.console.log(f"[cyan]Step: {step_name}[/cyan]")
+                for op in operations:
+                    if isinstance(op, dict):
+                        # Handle the case where the operation is a dictionary (e.g., for equijoin)
+                        op_name = list(op.keys())[0]
+                        op_details = op[op_name]
+                        self.console.log(f"  - {op_name}: {op_details}")
+                    else:
+                        self.console.log(f"  - {op}")
+                self.console.log("")  # Add a blank line between steps
+        else:
+            self.console.log("[yellow]No optimized operations found[/yellow]")
+
+    def should_optimize(self, step_name: str, op_name: str) -> Tuple[str, List[Dict[str, Any]], List[Dict[str, Any]], float]:
+        """
+        Determine if an operation should be optimized.
+        We do this by running the operations on a sample of the input data and checking if the output is correct.
+        """
+        self.console.rule("[bold cyan]Beginning Pipeline Optimization[/bold cyan]")
+        self.syntax_check()
+
+        self._insert_empty_resolve_operations()
+
+        for step in self.config["pipeline"]["steps"]:
+            self.captured_output.set_step(step.get("name"))
+            # Go through each operation in the step until we find the one we want to optimize
+            ops_run = []
+            op_name_to_object = {name: self.find_operation(name) for name in step["operations"]}
+            for op_idx, operation in enumerate(step["operations"]):
+                if isinstance(operation, dict):
+                    operation_name = list(operation.keys())[0]
+                    operation_config = operation[operation_name]
+                else:
+                    operation_name = operation
+                    operation_config = {}
+
+                op_object = self.find_operation(operation_name).copy()
+                op_object.update(operation_config)
+                op_object["name"] = operation_name
+
+                # Run the pipeline
+                sample_size = self.compute_sample_size(
+                    step.get("name"), step.get("operations"), op_object
+                )
+                input_data = self._run_partial_step(
+                    step, ops_run, sample_size, op_name_to_object
+                )
+                output_data = input_data
+
+                # If this is not the operation we want to optimize, just execute it and add to selectivities
+                if f"{step.get('name')}/{op_name}" != f"{step_name}/{op_name}" and op_object.get("empty", False):
+                    output_data = self._run_operation(op_object, input_data, is_build=True)
+                    self.selectivities[step.get("name")][op_name] = len(output_data) / len(input_data)
+                    ops_run.append(op_name)
+
+                # if this is the operation we want to optimize, invoke the optimizer's should_optimize method
+                else:
+                    if op_object.get("type") == "map" or op_object.get("type") == "filter":
+                        # Create instance of map optimizer
+                        map_optimizer = MapOptimizer(
+                            self,
+                            self.config,
+                            self.console,
+                            self.llm_client,
+                            self.max_threads,
+                            self._run_operation,
+                            timeout=self.timeout,
+                            is_filter=op_object.get("type") == "filter",
+                        )
+                        should_optimize_output, input_data, output_data = map_optimizer.should_optimize(op_object, input_data)
+                    elif op_object.get("type") == "reduce":
+                        reduce_optimizer = ReduceOptimizer(
+                            self.runner,
+                            self.config,
+                            self.console,
+                            self.llm_client,
+                            self.max_threads,
+                            self._run_operation,
+                        )
+                        should_optimize_output, input_data, output_data = reduce_optimizer.should_optimize(op_object, input_data)
+                    elif op_object.get("type") == "resolve":
+                        resolve_optimizer = JoinOptimizer(
+                            self.runner,
+                            self.config,
+                            op_object,
+                            self.console,
+                            self.llm_client,
+                            self.max_threads,
+                            target_recall=self.config.get("optimizer_config", {})
+                            .get("resolve", {})
+                            .get("target_recall", 0.95),
+                        )
+                        _, should_optimize_output = resolve_optimizer.should_optimize(input_data)
+
+                        # if should_optimize_output is empty, then we should move to the reduce operation
+                        if should_optimize_output == "":
+                            continue
+
+                    # Return the string and operation cost
+                    return should_optimize_output, input_data, output_data, self.operations_cost + self.llm_client.total_cost
+
+        # Should not get here
+        raise ValueError("No operation to optimize found")
+
+    def optimize(self) -> float:
+        """
+        Optimize the entire pipeline defined in the configuration.
+
+        This method is the main entry point for the optimization process. It iterates through
+        each step in the pipeline, optimizing from upstream to downstream, and constructs an
+        optimized version of the configuration.
+
+        The optimization process includes:
+        1. Iterating through each step in the pipeline, from upstream to downstream.
+        2. Optimizing each step using the _optimize_step method.
+        3. Updating the optimized configuration with the new operations and steps.
+        4. Saving the optimized configuration to a file.
+        5. Logging the total costs (agent cost, operations cost, and total cost).
+
+        Returns:
+            None
+
+        Side effects:
+        - Modifies self.optimized_config with the optimized pipeline and operations.
+        - Updates self.datasets with the results of each step.
+        - Calls _save_optimized_config to save the optimized configuration to a file.
+        - Logs cost information to the console.
+
+        Raises:
+            ValueError: If a step in the pipeline does not have a name.
+
+        Note:
+        - This method assumes that all necessary data and configurations are already
+          loaded and initialized in the Optimizer instance.
+        - The optimization process is performed step by step, from upstream to downstream,
+          with each step potentially depending on the results of previous steps.
+        """
+        self.console.rule("[bold cyan]Beginning Pipeline Optimization[/bold cyan]")
+
+        self.syntax_check()
+
+        self._insert_empty_resolve_operations()
+
+        # If resume is True, load the optimized operations from disk
+        if self.resume:
+            self._load_optimized_ops()
+
+        for step in self.config["pipeline"]["steps"]:
+            step_name = step.get("name")
+            if not step_name:
+                raise ValueError(
+                    "Step does not have a name. Each step must have a unique name."
+                )
+
+            optimized_step, step_operations, input_data = self._optimize_step(step)
+            old_op_names = [
+                op
+                for op in step["operations"]
+                if op not in optimized_step["operations"]
+            ]
+
+            # Remove all old_op_names from self.optimized_config["operations"]
+            self.optimized_config["operations"] = [
+                op
+                for op in self.optimized_config["operations"]
+                if op["name"] not in old_op_names
+            ]
+
+            for op in optimized_step["operations"]:
+                changed_op = False
+                for i, op_config in enumerate(self.optimized_config["operations"]):
+                    if op_config["name"] == op:
+                        self.optimized_config["operations"][i] = step_operations[op]
+                        changed_op = True
+                if not changed_op:
+                    self.optimized_config["operations"].append(step_operations[op])
+
+            self.optimized_config["pipeline"]["steps"] = [
+                step
+                for step in self.optimized_config["pipeline"]["steps"]
+                if step["name"] != step_name
+            ] + [optimized_step]
+
+            self.step_op_to_optimized_ops[step_name] = optimized_step["operations"]
+
+            step_hash = (
+                hashlib.md5(
+                    json.dumps(
+                        {
+                            "step": [
+                                s
+                                for s in self.optimized_config["pipeline"]["steps"]
+                                if s["name"] == step_name
+                            ][0],
+                            "operations": [
+                                self.find_operation(op, self.optimized_config)
+                                for op in optimized_step["operations"]
+                            ],
+                        }
+                    ).encode()
+                ).hexdigest()
+                + ".json"
+            )
+            # If the dataset already exists, skip the step
+            if step_hash in self.datasets:
+                continue
+
+            flush_cache(self.console)
+
+            if step_name in self.config.get("optimizer_config", {}).get(
+                "run_full_step", []
+            ):
+                # Run the entire step
+                input_data = self._run_partial_step(
+                    step,
+                    step_operations,
+                    float("inf"),  # TODO: FIX THIS
+                )
+                self.datasets[step_hash] = copy.deepcopy(input_data)
+            else:
+                self.datasets[step_hash] = copy.deepcopy(input_data)
+
+        self.console.log(
+            f"[bold]Total agent cost: ${self.llm_client.total_cost:.2f}[/bold]"
+        )
+        self.console.log(
+            f"[bold]Total operations cost: ${self.operations_cost:.2f}[/bold]"
+        )
+        self.console.log(
+            f"[bold]Total cost: ${self.llm_client.total_cost + self.operations_cost:.2f}[/bold]"
+        )
+
+        return self.llm_client.total_cost + self.operations_cost
+
+    def _run_partial_step(
+        self,
+        step: Dict[str, Any],
+        ops_to_run: List[str],
+        sample_size: int,
+        optimized_operations: Dict[str, Dict[str, Any]],
+    ) -> List[Dict[str, Any]]:
+        """
+        Execute a partial step of the pipeline on a sample of the input data.
+
+        This internal method runs a subset of operations for a given step on a sample
+        of the input data. It's used as part of the optimization process to evaluate
+        and optimize individual operations within a step.
+
+        Args:
+            step (Dict[str, Any]): The step configuration dictionary.
+            ops_to_run (List[str]): List of operation names to execute in this partial step.
+            sample_size (int): The number of items to include in the input sample.
+            optimized_operations (Dict[str, Dict[str, Any]]): Dictionary of optimized operations.
+
+        Returns:
+            List[Dict[str, Any]]: The output data after running the specified operations.
+
+        The method performs the following steps:
+        1. Retrieves a sample of the input data using _get_sample_data.
+        2. For equijoin operations, it loads both left and right datasets.
+        3. Iterates through the specified operations, running each on the input sample.
+        4. Returns the final output after all specified operations have been applied.
+
+        Note:
+        - The method handles both regular steps and equijoin steps differently.
+
+        Raises:
+            Any exceptions raised by _get_sample_data or _run_operation methods.
+        """
+        # Take the input data and run the operations in ops_to_run
+        # Return the output data
+        input_sample = self._get_sample_data(step.get("input"), None, sample_size)
+
+        if step.get("input") is None:
+            join_op_name = list(step.get("operations")[0].keys())[0]
+            # this is an equijoin step, load left and right datasets
+            left_data = self._get_sample_data(
+                step.get("operations")[0][join_op_name].get("left"), None, sample_size
+            )
+            right_data = self._get_sample_data(
+                step.get("operations")[0][join_op_name].get("right"), None, sample_size
+            )
+            input_sample = {"left": left_data, "right": right_data}
+
+        for op in ops_to_run:
+            op_object = optimized_operations[op]
+            if "name" not in op_object:
+                op_object["name"] = op
+
+            input_sample = self._run_operation(op_object, input_sample)
+        return input_sample
+
+    def _optimize_step(
+        self, step: Dict[str, Any]
+    ) -> Tuple[Dict[str, Any], List[Dict[str, Any]]]:
+        """
+        Optimize a single step in the pipeline.
+
+        This method takes a step configuration and optimizes each operation within it.
+        It handles different types of operations, including those that require optimization
+        and those that don't.
+
+        Args:
+            step (Dict[str, Any]): The configuration dictionary for the step to be optimized.
+
+        Returns:
+            Tuple[Dict[str, Any], List[Dict[str, Any]], List[Dict[str, Any]]]:
+                - The optimized step configuration.
+                - A list of optimized operations.
+                - The output data after running all operations in the step.
+
+        The method performs the following for each operation in the step:
+        1. Extracts the operation configuration.
+        2. Computes the appropriate sample size for the operation.
+        3. Runs the operation on a sample of the input data.
+        4. If the operation is optimizable and of a supported type, it calls the appropriate
+           optimization method (e.g., _optimize_map, _optimize_reduce).
+        5. If not optimizable or not supported, it runs the operation as-is.
+        6. Calculates and stores the selectivity of each operation.
+        7. Updates the list of optimized operations and their configurations.
+
+        The method uses rich console to provide status updates during the optimization process.
+
+        Note:
+        - This method is a key part of the overall optimization process, focusing on
+          individual steps in the pipeline.
+        - It relies on several helper methods like _run_partial_step, compute_sample_size,
+          and various _optimize_* methods for specific operation types.
+        - When optimizing an operation in the step, all previous operations are run on the
+          sample size needed for the current operation. This ensures that the input to the
+          operation being optimized is representative of what it would receive in the full pipeline.
+
+        Raises:
+            ValueError: If an unsupported operation type is encountered.
+        """
+        self.captured_output.set_step(step.get("name"))
+        optimized_operations = {}
+        optimized_operation_names = []
+        replacement_operations = {}  # List from old op name to new ops
+
+        for op_idx, operation in enumerate(step["operations"]):
+            if isinstance(operation, dict):
+                operation_name = list(operation.keys())[0]
+                operation_config = operation[operation_name]
+            else:
+                operation_name = operation
+                operation_config = {}
+
+            op_object = self.find_operation(operation_name).copy()
+            op_object.update(operation_config)
+            op_object["name"] = operation_name
+
+            # Run the pipeline
+            step_ops = []
+            for step_op in step.get("operations"):
+                if step_op in replacement_operations:
+                    step_ops.extend(replacement_operations[step_op])
+                else:
+                    step_ops.append(step_op)
+
+            # TODO: incorporate this into the optimizer to not run the most downstream operations
+            downstream_ops_exist = op_idx < len(step["operations"]) - 1
+
+            sample_size = self.compute_sample_size(
+                step.get("name"), step_ops, op_object
+            )
+            input_data = self._run_partial_step(
+                step, optimized_operation_names, sample_size, optimized_operations
+            )
+
+            if (
+                not op_object.get("optimize", False)  # Default don't optimize
+                or op_object.get("type") not in SUPPORTED_OPS
+            ):
+                # If optimize is False or operation type is not supported, just use the operation without optimization
+                output_data = self._run_operation(op_object, input_data)
+                optimized_operations[operation_name] = op_object
+                optimized_operation_names.append(operation_name)
+
+                selectivity = len(output_data) / len(input_data)
+
+                self.selectivities[step.get("name")][operation_name] = selectivity
+                self.samples_taken[step.get("name")][operation_name] = sample_size
+            else:
+                # Use rich console status to indicate optimization of the operation
+                with self.console.status(
+                    f"[bold blue]Optimizing operation: {operation_name} (Type: {op_object['type']})[/bold blue]"
+                ) as status:
+                    self.status = status
+
+                    # Print the number of elements in input_data
+                    self.console.rule(
+                        f"[yellow]Optimizing operation {operation_name} (Type: {op_object['type']})[/yellow]"
+                    )
+                    if op_object.get("type") == "equijoin":
+                        self.console.log(
+                            f"[yellow]  Sample size (left): {len(input_data['left'])}[/yellow]"
+                        )
+                        self.console.log(
+                            f"[yellow]  Sample size (right): {len(input_data['right'])}[/yellow]"
+                        )
+                    else:
+                        self.console.log(
+                            f"[yellow]  Sample size: {len(input_data)}[/yellow]"
+                        )
+
+                    # Run optimization
+                    for retry in range(
+                        self.config.get("optimizer_config", {}).get(
+                            "num_retries", NUM_OPTIMIZER_RETRIES
+                        )
+                    ):
+                        try:
+                            if op_object.get("type") == "map":
+                                optimized_ops = self._optimize_map(
+                                    op_object, input_data
+                                )
+                            elif op_object.get("type") == "filter":
+                                optimized_ops = self._optimize_map(
+                                    op_object, input_data, is_filter=True
+                                )
+                            elif op_object.get("type") == "reduce":
+                                optimized_ops = self._optimize_reduce(
+                                    op_object, input_data, status
+                                )
+                            elif op_object.get("type") == "resolve":
+                                optimized_ops = self._optimize_resolve(
+                                    op_object, input_data
+                                )
+                            elif op_object.get("type") == "equijoin":
+                                (
+                                    optimized_ops,
+                                    input_data,
+                                    new_left_name,
+                                    new_right_name,
+                                ) = self._optimize_equijoin(
+                                    op_object,
+                                    operation["left"],
+                                    operation["right"],
+                                    input_data["left"],
+                                    input_data["right"],
+                                    status,
+                                )
+                            else:
+                                raise ValueError(
+                                    f"Unsupported operation type: {op_object['type']}"
+                                )
+                            break  # If successful, break out of the retry loop
+                        except Exception as e:
+                            if (
+                                retry
+                                == self.config.get("optimizer_config", {}).get(
+                                    "num_retries", NUM_OPTIMIZER_RETRIES
+                                )
+                                - 1
+                            ):
+                                raise  # If this was the last retry, re-raise the exception
+                            self.console.log(
+                                f"Optimization attempt {retry + 1} failed. Retrying..."
+                            )
+
+                    if self.status:
+                        self.status.update(
+                            f"[bold blue]Running optimized operation to estimate selectivities: {operation_name}[/bold blue]"
+                        )
+
+                    for op in optimized_ops:
+                        op_name = op["name"]
+                        optimized_operations[op_name] = op
+                        if op.get("type") == "equijoin":
+                            optimized_operation_names.append(
+                                {
+                                    op_name: {
+                                        "left": new_left_name,
+                                        "right": new_right_name,
+                                    }
+                                }
+                            )
+                        else:
+                            optimized_operation_names.append(op_name)
+
+                        old_input_data_size = len(input_data)
+                        input_data = self._run_operation(op, input_data)
+                        new_input_data_size = len(input_data)
+                        selectivity = new_input_data_size / old_input_data_size
+                        self.selectivities[step.get("name")][op_name] = selectivity
+                        self.samples_taken[step.get("name")][op_name] = sample_size
+
+                    # Set replacement_operations
+                    replacement_operations[op_object["name"]] = [
+                        o["name"] for o in optimized_ops
+                    ]
+
+                    # Print new operator configs
+                    self.console.log("[bold green]New op configurations:[/bold green]")
+                    for op_name, op_config in optimized_operations.items():
+                        if op_name in [o["name"] for o in optimized_ops]:
+                            self.console.log(
+                                f"[cyan]{op_name}:[/cyan] {json.dumps(op_config, indent=2)}"
+                            )
+
+                    # Save the optimized operations to disk
+                    os.makedirs(self.optimized_ops_path, exist_ok=True)
+
+                    for original_op, replacement_ops in replacement_operations.items():
+                        optimized_ops_list = [
+                            (
+                                optimized_operations[op_name]
+                                if isinstance(op_name, str)
+                                else {
+                                    list(op_name.keys())[0]: optimized_operations[
+                                        list(op_name.keys())[0]
+                                    ]
+                                }
+                            )
+                            for op_name in replacement_ops
+                        ]
+
+                        # Save to disk
+                        optimized_op_file = os.path.join(
+                            self.optimized_ops_path, f"{original_op}.json"
+                        )
+                        with open(optimized_op_file, "w") as f:
+                            json.dump(optimized_ops_list, f, indent=2)
+
+                    self.console.log(
+                        f"[green]Saved optimized operations to {self.optimized_ops_path}[/green]"
+                    )
+                    self.status = None
+                    output_data = input_data
+
+        optimized_step = step.copy()
+        optimized_step["operations"] = optimized_operation_names
+        return optimized_step, optimized_operations, output_data
+
+    def _get_sample_data(
+        self, dataset_name: str, op_config: Optional[Dict[str, Any]], sample_size: int
+    ) -> List[Dict[str, Any]]:
+        """
+        Retrieve a sample of data from a specified dataset.
+
+        This method loads data from either a previously processed dataset or from a file,
+        and returns a sample of the data based on the given sample size and operation configuration.
+
+        Args:
+            dataset_name (str): The name of the dataset to sample from.
+            op_config (Optional[Dict[str, Any]]): The configuration of the operation to be performed.
+                                                  This is used to determine if special sampling is needed.
+            sample_size (int): The desired size of the sample. If set to float('inf'), all data is returned.
+
+        Returns:
+            List[Dict[str, Any]]: A list of dictionaries representing the sampled data.
+
+        Raises:
+            ValueError: If the dataset is not found or if the dataset type is unsupported.
+        """
+        if dataset_name is None:
+            return []
+
+        if any(
+            s["name"] == dataset_name
+            for s in self.optimized_config["pipeline"]["steps"]
+        ):
+            step = [
+                s
+                for s in self.optimized_config["pipeline"]["steps"]
+                if s["name"] == dataset_name
+            ][0]
+            name_hash = (
+                hashlib.md5(
+                    json.dumps(
+                        {
+                            "step": step,
+                            "operations": [
+                                self.find_operation(op) for op in step["operations"]
+                            ],
+                        }
+                    ).encode()
+                ).hexdigest()
+                + ".json"
+            )
+        else:
+            name_hash = None
+
+        if name_hash and name_hash in self.datasets:
+            data = self.datasets[name_hash]
+        else:
+            dataset_config = self.config["datasets"].get(dataset_name)
+            if dataset_config is None:
+                raise ValueError(
+                    f"Dataset '{dataset_name}' not found in config or previous steps."
+                )
+            dataset = Dataset(
+                runner=self,
+                type=dataset_config["type"],
+                path_or_data=dataset_config["path"],
+                parsing=dataset_config.get("parsing", []),
+                user_defined_parsing_tool_map=self.runner.parsing_tool_map,
+            )
+            data = dataset.load()
+
+        if sample_size == float("inf"):
+            return data
+
+        if op_config:
+            if op_config.get("type") == "reduce":
+                return self._get_reduce_sample(
+                    data, op_config.get("reduce_key"), sample_size
+                )
+
+        if not self.config.get("optimizer_config", {}).get("random_sample", False):
+            return data[:sample_size]
+
+        # Take the random 500 examples or all if less than 500
+        initial_data = random.sample(data, min(500, len(data)))
+
+        # Calculate counts for each example
+        char_counts = [len(str(item)) for item in initial_data]
+        total_counts = sum(char_counts)
+
+        # Calculate weights based on word counts
+        weights = [count / total_counts for count in char_counts]
+
+        # Perform weighted random sampling
+        return random.choices(
+            initial_data, weights=weights, k=min(sample_size, len(initial_data))
+        )
+
+    def _get_reduce_sample(
+        self, data: List[Dict[str, Any]], reduce_key: str, sample_size: int
+    ) -> List[Dict[str, Any]]:
+        """
+        Get a representative sample for a reduce operation.
+
+        This method creates a sample that preserves the distribution of groups in the data,
+        focusing on the top 5 largest groups. It also generates and prints a histogram of group sizes.
+
+        Args:
+            data (List[Dict[str, Any]]): The full dataset to sample from.
+            reduce_key (str): The key used for grouping in the reduce operation.
+            sample_size (int): The desired size of the sample.
+
+        Returns:
+            List[Dict[str, Any]]: A list of dictionaries representing the sampled data.
+        """
+        # Group data by reduce key
+        grouped_data = defaultdict(list)
+        for item in data:
+            grouped_data[item[reduce_key]].append(item)
+
+        # Sort groups by size in descending order
+        sorted_groups = sorted(
+            grouped_data.items(), key=lambda x: len(x[1]), reverse=True
+        )
+
+        sample = []
+
+        # Take the top 5 groups
+        top_5_groups = sorted_groups[:5]
+
+        # Calculate the total count of items in the top 5 groups
+        total_count = sum(len(items) for _, items in top_5_groups)
+
+        sample = []
+        for _, items in top_5_groups:
+            # Calculate the proportion of items to sample from this group
+            group_proportion = len(items) / total_count
+            group_sample_size = int(sample_size * group_proportion)
+
+            # Sample from the group
+            if not self.config.get("optimizer_config", {}).get("random_sample", False):
+                group_sample = items[:group_sample_size]
+            else:
+                group_sample = random.sample(
+                    items, min(group_sample_size, len(items))
+                )
+
+            sample.extend(group_sample)
+
+        # If we haven't reached the desired sample size, add more items randomly
+        if len(sample) < sample_size:
+            remaining_items = [
+                item
+                for _, items in top_5_groups
+                for item in items
+                if item not in sample
+            ]
+            additional_sample = random.sample(
+                remaining_items,
+                min(
+                    sample_size - len(sample), len(remaining_items)
+                ),
+            ) if self.config.get("optimizer_config", {}).get("random_sample", False) else remaining_items[:sample_size - len(sample)]
+            sample.extend(additional_sample)
+
+        # Create a histogram of group sizes
+        group_sizes = [len(items) for _, items in grouped_data.items()]
+        size_counts = Counter(group_sizes)
+
+        # Sort the sizes for a more readable output
+        sorted_sizes = sorted(size_counts.items())
+
+        # Print the histogram
+        self.console.log("\n[bold]Histogram of Group Sizes:[/bold]")
+        max_bar_width, max_count = 2, max(size_counts.values())
+        for size, count in sorted_sizes[:5]:
+            normalized_count = int(count / max_count * max_bar_width)
+            bar = "█" * normalized_count
+            self.console.log(f"{size:3d}: {bar} ({count})")
+        self.console.log("\n")
+
+        return sample
+
+    def _optimize_reduce(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        status: Status,
+    ) -> List[Dict[str, Any]]:
+        """
+        Optimize a reduce operation.
+
+        This method creates a ReduceOptimizer instance and uses it to optimize the reduce operation.
+
+        Args:
+            op_config (Dict[str, Any]): The configuration of the reduce operation.
+            input_data (List[Dict[str, Any]]): The input data for the reduce operation.
+            status (Status): The status object to update with the progress of the optimization.
+
+        Returns:
+            List[Dict[str, Any]]: The optimized operation configuration.
+        """
+        reduce_optimizer = ReduceOptimizer(
+            self.runner,
+            self.config,
+            self.console,
+            self.llm_client,
+            self.max_threads,
+            self._run_operation,
+            status=status,
+        )
+        optimized_ops, _, cost = reduce_optimizer.optimize(op_config, input_data)
+        self.operations_cost += cost
+        return optimized_ops
+
+    def _optimize_equijoin(
+        self,
+        op_config: Dict[str, Any],
+        left_name: str,
+        right_name: str,
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+        status: Status,
+    ) -> Tuple[List[Dict[str, Any]], Dict[str, List[Dict[str, Any]]], str, str]:
+        """
+        Optimize an equijoin operation.
+
+        This method creates a JoinOptimizer instance and uses it to optimize the equijoin operation.
+        It updates the operation cost and runs the optimized operation.
+        If the LLM suggests a map transformation, it will optimize the map operation as its own step, and then go back to optimize the equijoin operation.
+
+        Args:
+            op_config (Dict[str, Any]): The configuration of the equijoin operation.
+            left_name (str): The name of the left dataset.
+            right_name (str): The name of the right dataset.
+            left_data (List[Dict[str, Any]]): The left dataset for the join.
+            right_data (List[Dict[str, Any]]): The right dataset for the join.
+
+        Returns:
+            Tuple[List[Dict[str, Any]], Dict[str, List[Dict[str, Any]]], str, str]: The optimized operation configuration, the new left and right datasets, and the new left and right names.
+        """
+        max_iterations = 2
+        new_left_name = left_name
+        new_right_name = right_name
+        for _ in range(max_iterations):
+            join_optimizer = JoinOptimizer(
+                self.runner,
+                self.config,
+                op_config,
+                self.console,
+                self.llm_client,
+                self.max_threads,
+                target_recall=self.config.get("optimizer_config", {})
+                .get("equijoin", {})
+                .get("target_recall", 0.95),
+                estimated_selectivity=self.config.get("optimizer_config", {})
+                .get("equijoin", {})
+                .get("estimated_selectivity", None),
+                status=status,
+            )
+            optimized_config, cost, agent_results = join_optimizer.optimize_equijoin(
+                left_data, right_data
+            )
+            self.operations_cost += cost
+            # Update the operation config with the optimized values
+            op_config.update(optimized_config)
+
+            if not agent_results.get("optimize_map", False):
+                break  # Exit the loop if no more map optimizations are necessary
+
+            # Update the status to indicate we're optimizing a map operation
+            output_key = agent_results["output_key"]
+            if self.status:
+                self.status.update(
+                    f"Optimizing map operation for {output_key} extraction to help with the equijoin"
+                )
+            map_prompt = agent_results["map_prompt"]
+            dataset_to_transform = (
+                left_data
+                if agent_results["dataset_to_transform"] == "left"
+                else right_data
+            )
+
+            # Create a new step for the map operation
+            map_operation = {
+                "name": f"synthesized_{output_key}_extraction",
+                "type": "map",
+                "prompt": map_prompt,
+                "model": self.config.get("default_model", "gpt-4o-mini"),
+                "output": {"schema": {output_key: "string"}},
+                "optimize": False,
+            }
+
+            # Optimize the map operation
+            if map_operation["optimize"]:
+                dataset_to_transform_sample = random.sample(
+                    dataset_to_transform, self.sample_size_map.get("map")
+                ) if self.config.get("optimizer_config", {}).get("random_sample", False) else dataset_to_transform[:self.sample_size_map.get("map")]
+                optimized_map_operations = self._optimize_map(
+                    map_operation, dataset_to_transform_sample
+                )
+            else:
+                optimized_map_operations = [map_operation]
+
+            new_step = {
+                "name": f"synthesized_{output_key}_extraction",
+                "input": (
+                    left_name
+                    if agent_results["dataset_to_transform"] == "left"
+                    else right_name
+                ),
+                "operations": [mo["name"] for mo in optimized_map_operations],
+            }
+            if agent_results["dataset_to_transform"] == "left":
+                new_left_name = new_step["name"]
+            else:
+                new_right_name = new_step["name"]
+
+            for optimized_map_op in optimized_map_operations:
+                self.optimized_config["operations"].append(optimized_map_op)
+
+            self.optimized_config["pipeline"]["steps"].append(new_step)
+
+            # Now run the optimized map operation on the entire dataset_to_transform
+            for op in optimized_map_operations:
+                dataset_to_transform = self._run_operation(op, dataset_to_transform)
+
+            # Update the appropriate dataset for the next iteration
+            if agent_results["dataset_to_transform"] == "left":
+                left_data = dataset_to_transform
+            else:
+                right_data = dataset_to_transform
+
+            if self.status:
+                self.status.update(
+                    f"Optimizing equijoin operation with {output_key} extraction"
+                )
+
+        # Pop off "left" and "right" from the op_config
+        op_config.pop("left")
+        op_config.pop("right")
+        return (
+            [op_config],
+            {"left": left_data, "right": right_data},
+            new_left_name,
+            new_right_name,
+        )
+
+    def _optimize_map(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        is_filter: bool = False,
+    ) -> List[Dict[str, Any]]:
+        """
+        Optimize a map operation.
+
+        This method creates a MapOptimizer instance and uses it to optimize the map operation.
+
+        Args:
+            op_config (Dict[str, Any]): The configuration of the map operation.
+            input_data (List[Dict[str, Any]]): The input data for the map operation.
+            is_filter (bool, optional): If True, the operation is a filter operation. Defaults to False.
+
+        Returns:
+            List[Dict[str, Any]]: The optimized operation configuration.
+        """
+
+        map_optimizer = MapOptimizer(
+            self,
+            self.config,
+            self.console,
+            self.llm_client,
+            self.max_threads,
+            self._run_operation,
+            timeout=self.timeout,
+            is_filter=is_filter,
+        )
+        optimized_ops, _, cost = map_optimizer.optimize(op_config, input_data)
+        self.operations_cost += cost
+        return optimized_ops
+
+    def _optimize_resolve(
+        self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]
+    ) -> List[Dict[str, Any]]:
+        """
+        Optimize a resolve operation.
+
+        This method creates a JoinOptimizer instance and uses it to optimize the resolve operation.
+        It updates the operation cost and runs the optimized operation.
+
+        Args:
+            op_config (Dict[str, Any]): The configuration of the resolve operation.
+            input_data (List[Dict[str, Any]]): The input data for the resolve operation.
+
+        Returns:
+            List[Dict[str, Any]]: The optimized operation configuration.
+        """
+        optimized_config, cost = JoinOptimizer(
+            self.runner,
+            self.config,
+            op_config,
+            self.console,
+            self.llm_client,
+            self.max_threads,
+            target_recall=self.config.get("optimizer_config", {})
+            .get("resolve", {})
+            .get("target_recall", 0.95),
+        ).optimize_resolve(input_data)
+
+        if optimized_config.get("empty", False):
+            # Remove this operation from the pipeline and just return input data
+            return [], input_data
+
+        self.operations_cost += cost
+
+        # Update the operation config with the optimized values
+        op_config.update(optimized_config)
+
+        return [op_config]
+
+    def _run_operation(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        return_instance: bool = False,
+        is_build: bool = False,
+    ) -> Union[List[Dict[str, Any]], Tuple[List[Dict[str, Any]], BaseOperation]]:
+        """
+        Run a single operation based on its configuration.
+
+        This method creates an instance of the appropriate operation class and executes it.
+        It also updates the total operation cost.
+
+        Args:
+            op_config (Dict[str, Any]): The configuration of the operation to run.
+            input_data (List[Dict[str, Any]]): The input data for the operation.
+            return_instance (bool, optional): If True, return the operation instance along with the output data.
+
+        Returns:
+            Union[List[Dict[str, Any]], Tuple[List[Dict[str, Any]], BaseOperation]]:
+            If return_instance is False, returns the output data.
+            If return_instance is True, returns a tuple of the output data and the operation instance.
+        """
+        operation_class = get_operation(op_config["type"])
+
+        oc_kwargs = {
+            "runner": self.runner,
+            "config": op_config,
+            "default_model": self.config["default_model"],
+            "max_threads": self.max_threads,
+            "console": self.console,
+            "status": self.status,
+        }
+        operation_instance = operation_class(**oc_kwargs)
+        if op_config["type"] == "equijoin":
+            left_data = input_data["left"]
+            right_data = input_data["right"]
+            output_data, cost = operation_instance.execute(left_data, right_data)
+        elif op_config["type"] == "filter":
+            output_data, cost = operation_instance.execute(input_data, is_build)
+        else:
+            output_data, cost = operation_instance.execute(input_data)
+        self.operations_cost += cost
+        if return_instance:
+            return output_data, operation_instance
+        else:
+            return output_data
+
+    # Recursively resolve all anchors and aliases
+    @staticmethod
+    def resolve_anchors(data):
+        """
+        Recursively resolve all anchors and aliases in a nested data structure.
+
+        This static method traverses through dictionaries and lists, resolving any YAML anchors and aliases.
+
+        Args:
+            data: The data structure to resolve. Can be a dictionary, list, or any other type.
+
+        Returns:
+            The resolved data structure with all anchors and aliases replaced by their actual values.
+        """
+        if isinstance(data, dict):
+            return {k: Optimizer.resolve_anchors(v) for k, v in data.items()}
+        elif isinstance(data, list):
+            return [Optimizer.resolve_anchors(item) for item in data]
+        else:
+            return data
+
+    def clean_optimized_config(self):
+        """
+        Remove _intermediates from each operation in the optimized config
+        """
+        # Create a copy of the optimized config to modify
+        config_to_save = self.optimized_config.copy()
+
+        resolved_config = Optimizer.resolve_anchors(config_to_save)
+
+        # Remove _intermediates from each operation in resolved_config
+        if "operations" in resolved_config:
+            for op_config in resolved_config["operations"]:
+                if "_intermediates" in op_config:
+                    del op_config["_intermediates"]
+
+        return resolved_config
+
+    def save_optimized_config(self, optimized_config_path: str):
+        """
+        Save the optimized configuration to a YAML file.
+
+        This method creates a copy of the optimized configuration, resolves all anchors and aliases,
+        and saves it to a new YAML file. The new file name is based on the original file name with '_opt' appended.
+        """
+        resolved_config = self.clean_optimized_config()
+
+        with open(optimized_config_path, "w") as f:
+            yaml.safe_dump(resolved_config, f, default_flow_style=False, width=80)
+            self.console.log(
+                f"[green italic]💾 Optimized config saved to {optimized_config_path}[/green italic]"
+            )
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ __init__(runner, model='gpt-4o', resume=False, timeout=60) + +

+ + +
+ +

Initialize the Optimizer class.

+

This method sets up the optimizer with the given configuration file and parameters. +It loads the configuration, initializes the console for output, sets up the LLM client, +and prepares various attributes for optimization.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ yaml_file + + str + +
+

Path to the YAML configuration file.

+
+
+ required +
+ max_threads + + Optional[int] + +
+

Maximum number of threads to use for parallel processing. +If None, it will be set to (number of CPUs * 4).

+
+
+ required +
+ model + + str + +
+

The name of the language model to use. Defaults to "gpt-4o".

+
+
+ 'gpt-4o' +
+ resume + + bool + +
+

Whether to resume optimization from a previous run. Defaults to False.

+
+
+ False +
+ timeout + + int + +
+

Timeout in seconds for operations. Defaults to 60.

+
+
+ 60 +
+ + +

Attributes:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescription
yaml_file_path + str + +
+

Stores the path to the YAML file.

+
+
config + Dict + +
+

Stores the loaded configuration from the YAML file.

+
+
console + Console + +
+

Rich console for formatted output.

+
+
optimized_config + Dict + +
+

A copy of the original config to be optimized.

+
+
llm_client + LLMClient + +
+

Client for interacting with the language model.

+
+
max_threads + int + +
+

Maximum number of threads for parallel processing.

+
+
operations_cost + float + +
+

Tracks the total cost of operations.

+
+
timeout + int + +
+

Timeout for operations in seconds.

+
+
selectivities + defaultdict + +
+

Stores selectivity information for operations. +Selectivity is the ratio of output size to input size for an operation. +It's used to estimate how much data will flow through the pipeline after +each operation, which helps in optimizing subsequent operations and +determining appropriate sample sizes. For example, a selectivity of 0.5 +means an operation halves the size of its input data.

+
+
datasets + Dict + +
+

Stores loaded datasets.

+
+
+

The method also calls print_optimizer_config() to display the initial configuration.

+ +
+ Source code in docetl/builder.py +
 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
def __init__(
+    self,
+    runner: "DSLRunner",
+    model: str = "gpt-4o",
+    resume: bool = False,
+    timeout: int = 60,
+):
+    """
+    Initialize the Optimizer class.
+
+    This method sets up the optimizer with the given configuration file and parameters.
+    It loads the configuration, initializes the console for output, sets up the LLM client,
+    and prepares various attributes for optimization.
+
+    Args:
+        yaml_file (str): Path to the YAML configuration file.
+        max_threads (Optional[int]): Maximum number of threads to use for parallel processing.
+            If None, it will be set to (number of CPUs * 4).
+        model (str): The name of the language model to use. Defaults to "gpt-4o".
+        resume (bool): Whether to resume optimization from a previous run. Defaults to False.
+        timeout (int): Timeout in seconds for operations. Defaults to 60.
+
+    Attributes:
+        yaml_file_path (str): Stores the path to the YAML file.
+        config (Dict): Stores the loaded configuration from the YAML file.
+        console (Console): Rich console for formatted output.
+        optimized_config (Dict): A copy of the original config to be optimized.
+        llm_client (LLMClient): Client for interacting with the language model.
+        max_threads (int): Maximum number of threads for parallel processing.
+        operations_cost (float): Tracks the total cost of operations.
+        timeout (int): Timeout for operations in seconds.
+        selectivities (defaultdict): Stores selectivity information for operations.
+            Selectivity is the ratio of output size to input size for an operation.
+            It's used to estimate how much data will flow through the pipeline after
+            each operation, which helps in optimizing subsequent operations and
+            determining appropriate sample sizes. For example, a selectivity of 0.5
+            means an operation halves the size of its input data.
+        datasets (Dict): Stores loaded datasets.
+
+    The method also calls print_optimizer_config() to display the initial configuration.
+    """
+    self.config = runner.config
+    self.console = runner.console
+    self.max_threads = runner.max_threads
+
+    self.base_name = runner.base_name
+    self.yaml_file_suffix = runner.yaml_file_suffix
+    self.config = runner.config
+    self.runner = runner
+    self.status = runner.status
+
+    self.optimized_config = copy.deepcopy(self.config)
+    self.llm_client = LLMClient(model)
+    self.operations_cost = 0
+    self.timeout = timeout
+    self.selectivities = defaultdict(dict)
+    self.samples_taken = defaultdict(dict)
+    self.resume = resume
+    self.captured_output = CapturedOutput()
+
+    home_dir = os.environ.get("DOCETL_HOME_DIR", os.path.expanduser("~"))
+    cache_dir = os.path.join(home_dir, f".docetl/cache/{runner.yaml_file_suffix}")
+    os.makedirs(cache_dir, exist_ok=True)
+    self.datasets = DatasetOnDisk(dir=cache_dir, console=self.console)
+    self.optimized_ops_path = f"{cache_dir}/optimized_ops"
+
+    # Update sample size map
+    self.sample_size_map = SAMPLE_SIZE_MAP
+    if self.config.get("optimizer_config", {}).get("sample_sizes", {}):
+        self.sample_size_map.update(self.config["optimizer_config"]["sample_sizes"])
+
+    self.step_op_to_optimized_ops = {}
+
+    self.print_optimizer_config()
+
+
+
+ +
+ +
+ + +

+ clean_optimized_config() + +

+ + +
+ +

Remove _intermediates from each operation in the optimized config

+ +
+ Source code in docetl/builder.py +
1489
+1490
+1491
+1492
+1493
+1494
+1495
+1496
+1497
+1498
+1499
+1500
+1501
+1502
+1503
+1504
def clean_optimized_config(self):
+    """
+    Remove _intermediates from each operation in the optimized config
+    """
+    # Create a copy of the optimized config to modify
+    config_to_save = self.optimized_config.copy()
+
+    resolved_config = Optimizer.resolve_anchors(config_to_save)
+
+    # Remove _intermediates from each operation in resolved_config
+    if "operations" in resolved_config:
+        for op_config in resolved_config["operations"]:
+            if "_intermediates" in op_config:
+                del op_config["_intermediates"]
+
+    return resolved_config
+
+
+
+ +
+ +
+ + +

+ compute_sample_size(step_name, step_ops, op_config) + +

+ + +
+ +

Compute the sample size necessary for optimizing given operation based on upstream operations.

+

This method calculates an appropriate sample size for an operation, taking into +account the selectivities of upstream operations in the same step. It uses a +predefined sample size map (SAMPLE_SIZE_MAP) as a starting point.

+

For example, if we have a 'map' operation with a default sample size of 10, +and one upstream operation with a selectivity of 0.5, the computed sample size for the upstream operation would be: +10 / 0.5 = 20

+

This ensures that after applying the selectivity of the upstream operation, +we still have a representative sample size for the current operation.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ step_name + + str + +
+

The name of the current step in the pipeline.

+
+
+ required +
+ step_ops + + List[str] + +
+

A list of all operations in the current step.

+
+
+ required +
+ op_config + + Dict[str, Any] + +
+

The configuration dictionary for the current operation.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
Name TypeDescription
int + int + +
+

The computed sample size for the operation.

+
+
+

The method works as follows: +1. If there are no upstream operations, it returns the default sample size for the operation type. +2. Otherwise, it starts with the default sample size and adjusts it based on the selectivities + of upstream operations. +3. It iterates through upstream operations in reverse order, dividing the sample size by + each operation's selectivity. +4. The final result is rounded to the nearest integer.

+ + +

Raises:

+ + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If the selectivity for any upstream operation is not found.

+
+
+ + +
+ Note +
    +
  • The method assumes that selectivities for all upstream operations have been + previously computed and stored in self.selectivities.
  • +
  • The sample size is always at least 1, even after all adjustments.
  • +
+
+
+ Source code in docetl/builder.py +
214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
def compute_sample_size(
+    self,
+    step_name: str,
+    step_ops: List[str],
+    op_config: Dict[str, Any],
+) -> int:
+    """
+    Compute the sample size necessary for optimizing given operation based on upstream operations.
+
+    This method calculates an appropriate sample size for an operation, taking into
+    account the selectivities of upstream operations in the same step. It uses a
+    predefined sample size map (SAMPLE_SIZE_MAP) as a starting point.
+
+    For example, if we have a 'map' operation with a default sample size of 10,
+    and one upstream operation with a selectivity of 0.5, the computed sample size for the upstream operation would be:
+    10 / 0.5 = 20
+
+    This ensures that after applying the selectivity of the upstream operation,
+    we still have a representative sample size for the current operation.
+
+    Args:
+        step_name (str): The name of the current step in the pipeline.
+        step_ops (List[str]): A list of all operations in the current step.
+        op_config (Dict[str, Any]): The configuration dictionary for the current operation.
+
+    Returns:
+        int: The computed sample size for the operation.
+
+    The method works as follows:
+    1. If there are no upstream operations, it returns the default sample size for the operation type.
+    2. Otherwise, it starts with the default sample size and adjusts it based on the selectivities
+       of upstream operations.
+    3. It iterates through upstream operations in reverse order, dividing the sample size by
+       each operation's selectivity.
+    4. The final result is rounded to the nearest integer.
+
+    Raises:
+        ValueError: If the selectivity for any upstream operation is not found.
+
+    Note:
+        - The method assumes that selectivities for all upstream operations have been
+          previously computed and stored in self.selectivities.
+        - The sample size is always at least 1, even after all adjustments.
+    """
+    # If an equijoin, load the default. Equijoins are always first
+    if op_config.get("type") == "equijoin":
+        return SAMPLE_SIZE_MAP.get(op_config.get("type"))
+
+    # If there are no upstream operations, use the default sample_size
+    upstream_ops = []
+    for step_op in step_ops:
+        if step_op != op_config.get("name"):
+            if step_op in self.step_op_to_optimized_ops:
+                upstream_ops.extend(self.step_op_to_optimized_ops[step_op])
+            else:
+                upstream_ops.append(step_op)
+        else:
+            break
+
+    if len(upstream_ops) == 0:
+        return self.sample_size_map.get(op_config.get("type"), float("inf"))
+
+    # Otherwise, compute the sample size based on the upstream operations
+    sample_size = self.sample_size_map.get(op_config.get("type"), 100)
+    for op in reversed(upstream_ops):
+        # Use the selectivity of the upstream operation to compute the sample size
+        if op not in self.selectivities[step_name]:
+            raise ValueError(
+                f"Selectivity for operation {op} not found in selectivities. Other ops are {self.selectivities[step_name]}"
+            )
+
+        sample_size = sample_size / self.selectivities[step_name].get(op)
+
+    return int(math.ceil(sample_size))
+
+
+
+ +
+ +
+ + +

+ optimize() + +

+ + +
+ +

Optimize the entire pipeline defined in the configuration.

+

This method is the main entry point for the optimization process. It iterates through +each step in the pipeline, optimizing from upstream to downstream, and constructs an +optimized version of the configuration.

+

The optimization process includes: +1. Iterating through each step in the pipeline, from upstream to downstream. +2. Optimizing each step using the _optimize_step method. +3. Updating the optimized configuration with the new operations and steps. +4. Saving the optimized configuration to a file. +5. Logging the total costs (agent cost, operations cost, and total cost).

+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ float + +
+

None

+
+
+

Side effects: +- Modifies self.optimized_config with the optimized pipeline and operations. +- Updates self.datasets with the results of each step. +- Calls _save_optimized_config to save the optimized configuration to a file. +- Logs cost information to the console.

+ + +

Raises:

+ + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If a step in the pipeline does not have a name.

+
+
+

Note: +- This method assumes that all necessary data and configurations are already + loaded and initialized in the Optimizer instance. +- The optimization process is performed step by step, from upstream to downstream, + with each step potentially depending on the results of previous steps.

+ +
+ Source code in docetl/builder.py +
573
+574
+575
+576
+577
+578
+579
+580
+581
+582
+583
+584
+585
+586
+587
+588
+589
+590
+591
+592
+593
+594
+595
+596
+597
+598
+599
+600
+601
+602
+603
+604
+605
+606
+607
+608
+609
+610
+611
+612
+613
+614
+615
+616
+617
+618
+619
+620
+621
+622
+623
+624
+625
+626
+627
+628
+629
+630
+631
+632
+633
+634
+635
+636
+637
+638
+639
+640
+641
+642
+643
+644
+645
+646
+647
+648
+649
+650
+651
+652
+653
+654
+655
+656
+657
+658
+659
+660
+661
+662
+663
+664
+665
+666
+667
+668
+669
+670
+671
+672
+673
+674
+675
+676
+677
+678
+679
+680
+681
+682
+683
+684
+685
+686
+687
+688
+689
+690
+691
+692
+693
+694
+695
+696
+697
+698
+699
+700
+701
def optimize(self) -> float:
+    """
+    Optimize the entire pipeline defined in the configuration.
+
+    This method is the main entry point for the optimization process. It iterates through
+    each step in the pipeline, optimizing from upstream to downstream, and constructs an
+    optimized version of the configuration.
+
+    The optimization process includes:
+    1. Iterating through each step in the pipeline, from upstream to downstream.
+    2. Optimizing each step using the _optimize_step method.
+    3. Updating the optimized configuration with the new operations and steps.
+    4. Saving the optimized configuration to a file.
+    5. Logging the total costs (agent cost, operations cost, and total cost).
+
+    Returns:
+        None
+
+    Side effects:
+    - Modifies self.optimized_config with the optimized pipeline and operations.
+    - Updates self.datasets with the results of each step.
+    - Calls _save_optimized_config to save the optimized configuration to a file.
+    - Logs cost information to the console.
+
+    Raises:
+        ValueError: If a step in the pipeline does not have a name.
+
+    Note:
+    - This method assumes that all necessary data and configurations are already
+      loaded and initialized in the Optimizer instance.
+    - The optimization process is performed step by step, from upstream to downstream,
+      with each step potentially depending on the results of previous steps.
+    """
+    self.console.rule("[bold cyan]Beginning Pipeline Optimization[/bold cyan]")
+
+    self.syntax_check()
+
+    self._insert_empty_resolve_operations()
+
+    # If resume is True, load the optimized operations from disk
+    if self.resume:
+        self._load_optimized_ops()
+
+    for step in self.config["pipeline"]["steps"]:
+        step_name = step.get("name")
+        if not step_name:
+            raise ValueError(
+                "Step does not have a name. Each step must have a unique name."
+            )
+
+        optimized_step, step_operations, input_data = self._optimize_step(step)
+        old_op_names = [
+            op
+            for op in step["operations"]
+            if op not in optimized_step["operations"]
+        ]
+
+        # Remove all old_op_names from self.optimized_config["operations"]
+        self.optimized_config["operations"] = [
+            op
+            for op in self.optimized_config["operations"]
+            if op["name"] not in old_op_names
+        ]
+
+        for op in optimized_step["operations"]:
+            changed_op = False
+            for i, op_config in enumerate(self.optimized_config["operations"]):
+                if op_config["name"] == op:
+                    self.optimized_config["operations"][i] = step_operations[op]
+                    changed_op = True
+            if not changed_op:
+                self.optimized_config["operations"].append(step_operations[op])
+
+        self.optimized_config["pipeline"]["steps"] = [
+            step
+            for step in self.optimized_config["pipeline"]["steps"]
+            if step["name"] != step_name
+        ] + [optimized_step]
+
+        self.step_op_to_optimized_ops[step_name] = optimized_step["operations"]
+
+        step_hash = (
+            hashlib.md5(
+                json.dumps(
+                    {
+                        "step": [
+                            s
+                            for s in self.optimized_config["pipeline"]["steps"]
+                            if s["name"] == step_name
+                        ][0],
+                        "operations": [
+                            self.find_operation(op, self.optimized_config)
+                            for op in optimized_step["operations"]
+                        ],
+                    }
+                ).encode()
+            ).hexdigest()
+            + ".json"
+        )
+        # If the dataset already exists, skip the step
+        if step_hash in self.datasets:
+            continue
+
+        flush_cache(self.console)
+
+        if step_name in self.config.get("optimizer_config", {}).get(
+            "run_full_step", []
+        ):
+            # Run the entire step
+            input_data = self._run_partial_step(
+                step,
+                step_operations,
+                float("inf"),  # TODO: FIX THIS
+            )
+            self.datasets[step_hash] = copy.deepcopy(input_data)
+        else:
+            self.datasets[step_hash] = copy.deepcopy(input_data)
+
+    self.console.log(
+        f"[bold]Total agent cost: ${self.llm_client.total_cost:.2f}[/bold]"
+    )
+    self.console.log(
+        f"[bold]Total operations cost: ${self.operations_cost:.2f}[/bold]"
+    )
+    self.console.log(
+        f"[bold]Total cost: ${self.llm_client.total_cost + self.operations_cost:.2f}[/bold]"
+    )
+
+    return self.llm_client.total_cost + self.operations_cost
+
+
+
+ +
+ +
+ + +

+ print_optimizer_config() + +

+ + +
+ +

Print the current configuration of the optimizer.

+

This method uses the Rich console to display a formatted output of the optimizer's +configuration. It includes details such as the YAML file path, sample sizes for +different operation types, maximum number of threads, the language model being used, +and the timeout setting.

+

The output is color-coded and formatted for easy readability, with a header and +separator lines to clearly delineate the configuration information.

+ +
+ Source code in docetl/builder.py +
196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
def print_optimizer_config(self):
+    """
+    Print the current configuration of the optimizer.
+
+    This method uses the Rich console to display a formatted output of the optimizer's
+    configuration. It includes details such as the YAML file path, sample sizes for
+    different operation types, maximum number of threads, the language model being used,
+    and the timeout setting.
+
+    The output is color-coded and formatted for easy readability, with a header and
+    separator lines to clearly delineate the configuration information.
+    """
+    self.console.rule("[bold cyan]Optimizer Configuration[/bold cyan]")
+    self.console.log(f"[yellow]Sample Size:[/yellow] {self.sample_size_map}")
+    self.console.log(f"[yellow]Max Threads:[/yellow] {self.max_threads}")
+    self.console.log(f"[yellow]Model:[/yellow] {self.llm_client.model}")
+    self.console.log(f"[yellow]Timeout:[/yellow] {self.timeout} seconds")
+
+
+
+ +
+ +
+ + +

+ resolve_anchors(data) + + + staticmethod + + +

+ + +
+ +

Recursively resolve all anchors and aliases in a nested data structure.

+

This static method traverses through dictionaries and lists, resolving any YAML anchors and aliases.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ data + + +
+

The data structure to resolve. Can be a dictionary, list, or any other type.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ +
+

The resolved data structure with all anchors and aliases replaced by their actual values.

+
+
+ +
+ Source code in docetl/builder.py +
1469
+1470
+1471
+1472
+1473
+1474
+1475
+1476
+1477
+1478
+1479
+1480
+1481
+1482
+1483
+1484
+1485
+1486
+1487
@staticmethod
+def resolve_anchors(data):
+    """
+    Recursively resolve all anchors and aliases in a nested data structure.
+
+    This static method traverses through dictionaries and lists, resolving any YAML anchors and aliases.
+
+    Args:
+        data: The data structure to resolve. Can be a dictionary, list, or any other type.
+
+    Returns:
+        The resolved data structure with all anchors and aliases replaced by their actual values.
+    """
+    if isinstance(data, dict):
+        return {k: Optimizer.resolve_anchors(v) for k, v in data.items()}
+    elif isinstance(data, list):
+        return [Optimizer.resolve_anchors(item) for item in data]
+    else:
+        return data
+
+
+
+ +
+ +
+ + +

+ save_optimized_config(optimized_config_path) + +

+ + +
+ +

Save the optimized configuration to a YAML file.

+

This method creates a copy of the optimized configuration, resolves all anchors and aliases, +and saves it to a new YAML file. The new file name is based on the original file name with '_opt' appended.

+ +
+ Source code in docetl/builder.py +
1506
+1507
+1508
+1509
+1510
+1511
+1512
+1513
+1514
+1515
+1516
+1517
+1518
+1519
def save_optimized_config(self, optimized_config_path: str):
+    """
+    Save the optimized configuration to a YAML file.
+
+    This method creates a copy of the optimized configuration, resolves all anchors and aliases,
+    and saves it to a new YAML file. The new file name is based on the original file name with '_opt' appended.
+    """
+    resolved_config = self.clean_optimized_config()
+
+    with open(optimized_config_path, "w") as f:
+        yaml.safe_dump(resolved_config, f, default_flow_style=False, width=80)
+        self.console.log(
+            f"[green italic]💾 Optimized config saved to {optimized_config_path}[/green italic]"
+        )
+
+
+
+ +
+ +
+ + +

+ should_optimize(step_name, op_name) + +

+ + +
+ +

Determine if an operation should be optimized. +We do this by running the operations on a sample of the input data and checking if the output is correct.

+ +
+ Source code in docetl/builder.py +
482
+483
+484
+485
+486
+487
+488
+489
+490
+491
+492
+493
+494
+495
+496
+497
+498
+499
+500
+501
+502
+503
+504
+505
+506
+507
+508
+509
+510
+511
+512
+513
+514
+515
+516
+517
+518
+519
+520
+521
+522
+523
+524
+525
+526
+527
+528
+529
+530
+531
+532
+533
+534
+535
+536
+537
+538
+539
+540
+541
+542
+543
+544
+545
+546
+547
+548
+549
+550
+551
+552
+553
+554
+555
+556
+557
+558
+559
+560
+561
+562
+563
+564
+565
+566
+567
+568
+569
+570
+571
def should_optimize(self, step_name: str, op_name: str) -> Tuple[str, List[Dict[str, Any]], List[Dict[str, Any]], float]:
+    """
+    Determine if an operation should be optimized.
+    We do this by running the operations on a sample of the input data and checking if the output is correct.
+    """
+    self.console.rule("[bold cyan]Beginning Pipeline Optimization[/bold cyan]")
+    self.syntax_check()
+
+    self._insert_empty_resolve_operations()
+
+    for step in self.config["pipeline"]["steps"]:
+        self.captured_output.set_step(step.get("name"))
+        # Go through each operation in the step until we find the one we want to optimize
+        ops_run = []
+        op_name_to_object = {name: self.find_operation(name) for name in step["operations"]}
+        for op_idx, operation in enumerate(step["operations"]):
+            if isinstance(operation, dict):
+                operation_name = list(operation.keys())[0]
+                operation_config = operation[operation_name]
+            else:
+                operation_name = operation
+                operation_config = {}
+
+            op_object = self.find_operation(operation_name).copy()
+            op_object.update(operation_config)
+            op_object["name"] = operation_name
+
+            # Run the pipeline
+            sample_size = self.compute_sample_size(
+                step.get("name"), step.get("operations"), op_object
+            )
+            input_data = self._run_partial_step(
+                step, ops_run, sample_size, op_name_to_object
+            )
+            output_data = input_data
+
+            # If this is not the operation we want to optimize, just execute it and add to selectivities
+            if f"{step.get('name')}/{op_name}" != f"{step_name}/{op_name}" and op_object.get("empty", False):
+                output_data = self._run_operation(op_object, input_data, is_build=True)
+                self.selectivities[step.get("name")][op_name] = len(output_data) / len(input_data)
+                ops_run.append(op_name)
+
+            # if this is the operation we want to optimize, invoke the optimizer's should_optimize method
+            else:
+                if op_object.get("type") == "map" or op_object.get("type") == "filter":
+                    # Create instance of map optimizer
+                    map_optimizer = MapOptimizer(
+                        self,
+                        self.config,
+                        self.console,
+                        self.llm_client,
+                        self.max_threads,
+                        self._run_operation,
+                        timeout=self.timeout,
+                        is_filter=op_object.get("type") == "filter",
+                    )
+                    should_optimize_output, input_data, output_data = map_optimizer.should_optimize(op_object, input_data)
+                elif op_object.get("type") == "reduce":
+                    reduce_optimizer = ReduceOptimizer(
+                        self.runner,
+                        self.config,
+                        self.console,
+                        self.llm_client,
+                        self.max_threads,
+                        self._run_operation,
+                    )
+                    should_optimize_output, input_data, output_data = reduce_optimizer.should_optimize(op_object, input_data)
+                elif op_object.get("type") == "resolve":
+                    resolve_optimizer = JoinOptimizer(
+                        self.runner,
+                        self.config,
+                        op_object,
+                        self.console,
+                        self.llm_client,
+                        self.max_threads,
+                        target_recall=self.config.get("optimizer_config", {})
+                        .get("resolve", {})
+                        .get("target_recall", 0.95),
+                    )
+                    _, should_optimize_output = resolve_optimizer.should_optimize(input_data)
+
+                    # if should_optimize_output is empty, then we should move to the reduce operation
+                    if should_optimize_output == "":
+                        continue
+
+                # Return the string and operation cost
+                return should_optimize_output, input_data, output_data, self.operations_cost + self.llm_client.total_cost
+
+    # Should not get here
+    raise ValueError("No operation to optimize found")
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Perform a syntax check on all operations defined in the configuration.

+

This method validates each operation by attempting to instantiate it. +If any operation fails to instantiate, a ValueError is raised.

+ + +

Raises:

+ + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If any operation fails the syntax check.

+
+
+ +
+ Source code in docetl/builder.py +
166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
def syntax_check(self):
+    """
+    Perform a syntax check on all operations defined in the configuration.
+
+    This method validates each operation by attempting to instantiate it.
+    If any operation fails to instantiate, a ValueError is raised.
+
+    Raises:
+        ValueError: If any operation fails the syntax check.
+    """
+    for operation_config in self.config["operations"]:
+        operation = operation_config["name"]
+        operation_type = operation_config["type"]
+
+        try:
+            operation_class = get_operation(operation_type)
+            operation_class(
+                self.runner,
+                operation_config,
+                self.config.get("default_model", "gpt-4o-mini"),
+                self.max_threads,
+                console=self.console,
+            )
+        except Exception as e:
+            raise ValueError(
+                f"Syntax check failed for operation '{operation}': {str(e)}"
+            )
+
+    self.console.log("[green]Syntax check passed for all operations.[/green]")
+
+
+
+ +
+ + + +
+ +
+ +
+ + + + + + + + + + + + + +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/api-reference/operations/index.html b/api-reference/operations/index.html new file mode 100644 index 00000000..f498acbc --- /dev/null +++ b/api-reference/operations/index.html @@ -0,0 +1,16517 @@ + + + + + + + + + + + + + + + + + + + + + + + docetl.operations - docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + +
+
+ +
+
+ + + +
+
+ + + + + + + +

LLM-Powered Operators

+ + +
+ + + +

+ docetl.operations.map.MapOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + + + + + + +
+ Source code in docetl/operations/map.py +
 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
class MapOperation(BaseOperation):
+    class schema(BaseOperation.schema):
+        type: str = "map"
+        output: Optional[Dict[str, Any]] = None
+        prompt: Optional[str] = None
+        model: Optional[str] = None
+        optimize: Optional[bool] = None
+        recursively_optimize: Optional[bool] = None
+        sample_size: Optional[int] = None
+        tools: Optional[List[Dict[str, Any]]] = (
+            None  # FIXME: Why isn't this using the Tool data class so validation works automatically?
+        )
+        validation_rules: Optional[List[str]] = Field(None, alias="validate")
+        num_retries_on_validate_failure: Optional[int] = None
+        gleaning: Optional[Dict[str, Any]] = None
+        drop_keys: Optional[List[str]] = None
+        timeout: Optional[int] = None
+        enable_observability: bool = False
+        batch_size: Optional[int] = None
+        clustering_method: Optional[str] = None
+        batch_prompt: Optional[str] = None
+        litellm_completion_kwargs: Dict[str, Any] = Field(default_factory=dict)
+        @field_validator("drop_keys")
+        def validate_drop_keys(cls, v):
+            if isinstance(v, str):
+                return [v]
+            return v
+
+    def __init__(
+        self,
+        *args,
+        **kwargs,
+    ):
+        super().__init__(*args, **kwargs)
+        self.max_batch_size: int = self.config.get(
+            "max_batch_size", kwargs.get("max_batch_size", None)
+        )
+        self.clustering_method = "random"
+
+    def syntax_check(self) -> None:
+        """
+            Checks the configuration of the MapOperation for required keys and valid structure.
+
+        Raises:
+            ValueError: If required keys are missing or invalid in the configuration.
+            TypeError: If configuration values have incorrect types.
+        """
+        config = self.schema(**self.config)
+
+        if config.drop_keys:
+            if any(not isinstance(key, str) for key in config.drop_keys):
+                raise TypeError("All items in 'drop_keys' must be strings")
+        elif not (config.prompt and config.output):
+            raise ValueError(
+                "If 'drop_keys' is not specified, both 'prompt' and 'output' must be present in the configuration"
+            )
+
+        if config.batch_prompt:
+            try:
+                template = Template(config.batch_prompt)
+                # Test render with a minimal inputs list to validate template
+                template.render(inputs=[{}])
+            except Exception as e:
+                raise ValueError(
+                    f"Invalid Jinja2 template in 'batch_prompt' or missing required 'inputs' variable: {str(e)}"
+                ) from e
+
+        if config.prompt or config.output:
+            for key in ["prompt", "output"]:
+                if not getattr(config, key):
+                    raise ValueError(
+                        f"Missing required key '{key}' in MapOperation configuration"
+                    )
+
+            if config.output and not config.output["schema"]:
+                raise ValueError("Missing 'schema' in 'output' configuration")
+
+            if config.prompt:
+                try:
+                    Template(config.prompt)
+                except Exception as e:
+                    raise ValueError(
+                        f"Invalid Jinja2 template in 'prompt': {str(e)}"
+                    ) from e
+
+            if config.model and not isinstance(config.model, str):
+                raise TypeError("'model' in configuration must be a string")
+
+            if config.tools:
+                for tool in config.tools:
+                    try:
+                        tool_obj = Tool(**tool)
+                    except Exception as e:
+                        raise TypeError("Tool must be a dictionary")
+
+                    if not (tool_obj.code and tool_obj.function):
+                        raise ValueError(
+                            "Tool is missing required 'code' or 'function' key"
+                        )
+
+                    if not isinstance(tool_obj.function, ToolFunction):
+                        raise TypeError("'function' in tool must be a dictionary")
+
+                    for key in ["name", "description", "parameters"]:
+                        if not getattr(tool_obj.function, key):
+                            raise ValueError(
+                                f"Tool is missing required '{key}' in 'function'"
+                            )
+
+            self.gleaning_check()
+
+
+    def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+        """
+        Executes the map operation on the provided input data.
+
+        Args:
+            input_data (List[Dict]): The input data to process.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.
+
+        This method performs the following steps:
+        1. If a prompt is specified, it processes each input item using the specified prompt and LLM model
+        2. Applies gleaning if configured
+        3. Validates the output
+        4. If drop_keys is specified, it drops the specified keys from each document
+        5. Aggregates results and calculates total cost
+
+        The method uses parallel processing to improve performance.
+        """
+        # Check if there's no prompt and only drop_keys
+        if "prompt" not in self.config and "drop_keys" in self.config:
+            # If only drop_keys is specified, simply drop the keys and return
+            dropped_results = []
+            for item in input_data:
+                new_item = {
+                    k: v for k, v in item.items() if k not in self.config["drop_keys"]
+                }
+                dropped_results.append(new_item)
+            return dropped_results, 0.0  # Return the modified data with no cost
+
+        if self.status:
+            self.status.stop()
+
+        def _process_map_item(item: Dict, initial_result: Optional[Dict] = None) -> Tuple[Optional[Dict], float]:
+            prompt_template = Template(self.config["prompt"])
+            prompt = prompt_template.render(input=item)
+
+            def validation_fn(response: Union[Dict[str, Any], ModelResponse]):
+                output = self.runner.api.parse_llm_response(
+                    response,
+                    schema=self.config["output"]["schema"],
+                    tools=self.config.get("tools", None),
+                    manually_fix_errors=self.manually_fix_errors,
+                )[0] if isinstance(response, ModelResponse) else response
+                for key, value in item.items():
+                    if key not in self.config["output"]["schema"]:
+                        output[key] = value
+                if self.runner.api.validate_output(self.config, output, self.console):
+                    return output, True
+                return output, False
+
+            self.runner.rate_limiter.try_acquire("call", weight=1)
+            llm_result = self.runner.api.call_llm(
+                self.config.get("model", self.default_model),
+                "map",
+                [{"role": "user", "content": prompt}],
+                self.config["output"]["schema"],
+                tools=self.config.get("tools", None),
+                scratchpad=None,
+                timeout_seconds=self.config.get("timeout", 120),
+                max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+                validation_config=(
+                    {
+                        "num_retries": self.num_retries_on_validate_failure,
+                        "val_rule": self.config.get("validate", []),
+                        "validation_fn": validation_fn,
+                    }
+                    if self.config.get("validate", None)
+                    else None
+                ),
+                gleaning_config=self.config.get("gleaning", None),
+                verbose=self.config.get("verbose", False),
+                bypass_cache=self.config.get("bypass_cache", False),
+                initial_result=initial_result,
+                litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+            )
+
+            if llm_result.validated:
+                # Parse the response
+                if isinstance(llm_result.response, ModelResponse):
+                    output = self.runner.api.parse_llm_response(
+                        llm_result.response,
+                        schema=self.config["output"]["schema"],
+                        tools=self.config.get("tools", None),
+                        manually_fix_errors=self.manually_fix_errors,
+                    )[0]
+                else:
+                    output = llm_result.response
+
+
+                # Augment the output with the original item
+                output = {**item, **output}
+                if self.config.get("enable_observability", False):
+                    output[f"_observability_{self.config['name']}"] = {"prompt": prompt}
+                return output, llm_result.total_cost
+
+            return None, llm_result.total_cost
+
+         # If there's a batch prompt, let's use that
+        def _process_map_batch(items: List[Dict]) -> Tuple[List[Dict], float]:
+            total_cost = 0
+            if len(items) > 1 and self.config.get("batch_prompt", None):
+                batch_prompt_template = Template(self.config["batch_prompt"])
+                batch_prompt = batch_prompt_template.render(inputs=items)
+
+                # Issue the batch call
+                llm_result = self.runner.api.call_llm_batch(
+                    self.config.get("model", self.default_model),
+                    "batch map",
+                    [{"role": "user", "content": batch_prompt}],
+                    self.config["output"]["schema"],
+                    verbose=self.config.get("verbose", False),
+                    timeout_seconds=self.config.get("timeout", 120),
+                    max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+                    bypass_cache=self.config.get("bypass_cache", False),
+                    litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+                )
+                total_cost += llm_result.total_cost
+
+                # Parse the LLM response
+                parsed_output = self.runner.api.parse_llm_response(llm_result.response, self.config["output"]["schema"])[0].get("results", [])
+                items_and_outputs = [(item, parsed_output[idx] if idx < len(parsed_output) else None) for idx, item in enumerate(items)]
+            else:
+                items_and_outputs = [(item, None) for item in items]
+
+            # Run _process_map_item for each item 
+            all_results = []
+            if len(items_and_outputs) > 1:
+                with ThreadPoolExecutor(max_workers=self.max_batch_size) as executor:
+                    futures = [executor.submit(_process_map_item, items_and_outputs[i][0], items_and_outputs[i][1]) for i in range(len(items_and_outputs))]
+                    for i in range(len(futures)):
+                        result, item_cost = futures[i].result()
+                        if result is not None:
+                            all_results.append(result)
+                        total_cost += item_cost
+            else:
+                result, item_cost = _process_map_item(items_and_outputs[0][0], items_and_outputs[0][1])
+                if result is not None:
+                    all_results.append(result)
+                total_cost += item_cost
+
+            # Return items and cost
+            return all_results, total_cost
+
+        with ThreadPoolExecutor(max_workers=self.max_batch_size) as executor:
+            batch_size = self.max_batch_size if self.max_batch_size is not None else 1
+            futures = []
+            for i in range(0, len(input_data), batch_size):
+                batch = input_data[i:i + batch_size]
+                futures.append(executor.submit(_process_map_batch, batch))
+            results = []
+            total_cost = 0
+            pbar = RichLoopBar(
+                range(len(futures)),
+                desc=f"Processing {self.config['name']} (map) on all documents",
+                console=self.console,
+            )
+            for i in pbar:
+                result_list, item_cost = futures[i].result()
+                if result_list:
+                    if "drop_keys" in self.config:
+                        result_list = [{
+                            k: v
+                            for k, v in result.items()
+                            if k not in self.config["drop_keys"]
+                        } for result in result_list]
+                    results.extend(result_list)
+                total_cost += item_cost
+
+        if self.status:
+            self.status.start()
+
+        return results, total_cost
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ execute(input_data) + +

+ + +
+ +

Executes the map operation on the provided input data.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ input_data + + List[Dict] + +
+

The input data to process.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[List[Dict], float] + +
+

Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.

+
+
+

This method performs the following steps: +1. If a prompt is specified, it processes each input item using the specified prompt and LLM model +2. Applies gleaning if configured +3. Validates the output +4. If drop_keys is specified, it drops the specified keys from each document +5. Aggregates results and calculates total cost

+

The method uses parallel processing to improve performance.

+ +
+ Source code in docetl/operations/map.py +
144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+    """
+    Executes the map operation on the provided input data.
+
+    Args:
+        input_data (List[Dict]): The input data to process.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.
+
+    This method performs the following steps:
+    1. If a prompt is specified, it processes each input item using the specified prompt and LLM model
+    2. Applies gleaning if configured
+    3. Validates the output
+    4. If drop_keys is specified, it drops the specified keys from each document
+    5. Aggregates results and calculates total cost
+
+    The method uses parallel processing to improve performance.
+    """
+    # Check if there's no prompt and only drop_keys
+    if "prompt" not in self.config and "drop_keys" in self.config:
+        # If only drop_keys is specified, simply drop the keys and return
+        dropped_results = []
+        for item in input_data:
+            new_item = {
+                k: v for k, v in item.items() if k not in self.config["drop_keys"]
+            }
+            dropped_results.append(new_item)
+        return dropped_results, 0.0  # Return the modified data with no cost
+
+    if self.status:
+        self.status.stop()
+
+    def _process_map_item(item: Dict, initial_result: Optional[Dict] = None) -> Tuple[Optional[Dict], float]:
+        prompt_template = Template(self.config["prompt"])
+        prompt = prompt_template.render(input=item)
+
+        def validation_fn(response: Union[Dict[str, Any], ModelResponse]):
+            output = self.runner.api.parse_llm_response(
+                response,
+                schema=self.config["output"]["schema"],
+                tools=self.config.get("tools", None),
+                manually_fix_errors=self.manually_fix_errors,
+            )[0] if isinstance(response, ModelResponse) else response
+            for key, value in item.items():
+                if key not in self.config["output"]["schema"]:
+                    output[key] = value
+            if self.runner.api.validate_output(self.config, output, self.console):
+                return output, True
+            return output, False
+
+        self.runner.rate_limiter.try_acquire("call", weight=1)
+        llm_result = self.runner.api.call_llm(
+            self.config.get("model", self.default_model),
+            "map",
+            [{"role": "user", "content": prompt}],
+            self.config["output"]["schema"],
+            tools=self.config.get("tools", None),
+            scratchpad=None,
+            timeout_seconds=self.config.get("timeout", 120),
+            max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+            validation_config=(
+                {
+                    "num_retries": self.num_retries_on_validate_failure,
+                    "val_rule": self.config.get("validate", []),
+                    "validation_fn": validation_fn,
+                }
+                if self.config.get("validate", None)
+                else None
+            ),
+            gleaning_config=self.config.get("gleaning", None),
+            verbose=self.config.get("verbose", False),
+            bypass_cache=self.config.get("bypass_cache", False),
+            initial_result=initial_result,
+            litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+        )
+
+        if llm_result.validated:
+            # Parse the response
+            if isinstance(llm_result.response, ModelResponse):
+                output = self.runner.api.parse_llm_response(
+                    llm_result.response,
+                    schema=self.config["output"]["schema"],
+                    tools=self.config.get("tools", None),
+                    manually_fix_errors=self.manually_fix_errors,
+                )[0]
+            else:
+                output = llm_result.response
+
+
+            # Augment the output with the original item
+            output = {**item, **output}
+            if self.config.get("enable_observability", False):
+                output[f"_observability_{self.config['name']}"] = {"prompt": prompt}
+            return output, llm_result.total_cost
+
+        return None, llm_result.total_cost
+
+     # If there's a batch prompt, let's use that
+    def _process_map_batch(items: List[Dict]) -> Tuple[List[Dict], float]:
+        total_cost = 0
+        if len(items) > 1 and self.config.get("batch_prompt", None):
+            batch_prompt_template = Template(self.config["batch_prompt"])
+            batch_prompt = batch_prompt_template.render(inputs=items)
+
+            # Issue the batch call
+            llm_result = self.runner.api.call_llm_batch(
+                self.config.get("model", self.default_model),
+                "batch map",
+                [{"role": "user", "content": batch_prompt}],
+                self.config["output"]["schema"],
+                verbose=self.config.get("verbose", False),
+                timeout_seconds=self.config.get("timeout", 120),
+                max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+                bypass_cache=self.config.get("bypass_cache", False),
+                litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+            )
+            total_cost += llm_result.total_cost
+
+            # Parse the LLM response
+            parsed_output = self.runner.api.parse_llm_response(llm_result.response, self.config["output"]["schema"])[0].get("results", [])
+            items_and_outputs = [(item, parsed_output[idx] if idx < len(parsed_output) else None) for idx, item in enumerate(items)]
+        else:
+            items_and_outputs = [(item, None) for item in items]
+
+        # Run _process_map_item for each item 
+        all_results = []
+        if len(items_and_outputs) > 1:
+            with ThreadPoolExecutor(max_workers=self.max_batch_size) as executor:
+                futures = [executor.submit(_process_map_item, items_and_outputs[i][0], items_and_outputs[i][1]) for i in range(len(items_and_outputs))]
+                for i in range(len(futures)):
+                    result, item_cost = futures[i].result()
+                    if result is not None:
+                        all_results.append(result)
+                    total_cost += item_cost
+        else:
+            result, item_cost = _process_map_item(items_and_outputs[0][0], items_and_outputs[0][1])
+            if result is not None:
+                all_results.append(result)
+            total_cost += item_cost
+
+        # Return items and cost
+        return all_results, total_cost
+
+    with ThreadPoolExecutor(max_workers=self.max_batch_size) as executor:
+        batch_size = self.max_batch_size if self.max_batch_size is not None else 1
+        futures = []
+        for i in range(0, len(input_data), batch_size):
+            batch = input_data[i:i + batch_size]
+            futures.append(executor.submit(_process_map_batch, batch))
+        results = []
+        total_cost = 0
+        pbar = RichLoopBar(
+            range(len(futures)),
+            desc=f"Processing {self.config['name']} (map) on all documents",
+            console=self.console,
+        )
+        for i in pbar:
+            result_list, item_cost = futures[i].result()
+            if result_list:
+                if "drop_keys" in self.config:
+                    result_list = [{
+                        k: v
+                        for k, v in result.items()
+                        if k not in self.config["drop_keys"]
+                    } for result in result_list]
+                results.extend(result_list)
+            total_cost += item_cost
+
+    if self.status:
+        self.status.start()
+
+    return results, total_cost
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +
Checks the configuration of the MapOperation for required keys and valid structure.
+
+ + +

Raises:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If required keys are missing or invalid in the configuration.

+
+
+ TypeError + +
+

If configuration values have incorrect types.

+
+
+ +
+ Source code in docetl/operations/map.py +
 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
def syntax_check(self) -> None:
+    """
+        Checks the configuration of the MapOperation for required keys and valid structure.
+
+    Raises:
+        ValueError: If required keys are missing or invalid in the configuration.
+        TypeError: If configuration values have incorrect types.
+    """
+    config = self.schema(**self.config)
+
+    if config.drop_keys:
+        if any(not isinstance(key, str) for key in config.drop_keys):
+            raise TypeError("All items in 'drop_keys' must be strings")
+    elif not (config.prompt and config.output):
+        raise ValueError(
+            "If 'drop_keys' is not specified, both 'prompt' and 'output' must be present in the configuration"
+        )
+
+    if config.batch_prompt:
+        try:
+            template = Template(config.batch_prompt)
+            # Test render with a minimal inputs list to validate template
+            template.render(inputs=[{}])
+        except Exception as e:
+            raise ValueError(
+                f"Invalid Jinja2 template in 'batch_prompt' or missing required 'inputs' variable: {str(e)}"
+            ) from e
+
+    if config.prompt or config.output:
+        for key in ["prompt", "output"]:
+            if not getattr(config, key):
+                raise ValueError(
+                    f"Missing required key '{key}' in MapOperation configuration"
+                )
+
+        if config.output and not config.output["schema"]:
+            raise ValueError("Missing 'schema' in 'output' configuration")
+
+        if config.prompt:
+            try:
+                Template(config.prompt)
+            except Exception as e:
+                raise ValueError(
+                    f"Invalid Jinja2 template in 'prompt': {str(e)}"
+                ) from e
+
+        if config.model and not isinstance(config.model, str):
+            raise TypeError("'model' in configuration must be a string")
+
+        if config.tools:
+            for tool in config.tools:
+                try:
+                    tool_obj = Tool(**tool)
+                except Exception as e:
+                    raise TypeError("Tool must be a dictionary")
+
+                if not (tool_obj.code and tool_obj.function):
+                    raise ValueError(
+                        "Tool is missing required 'code' or 'function' key"
+                    )
+
+                if not isinstance(tool_obj.function, ToolFunction):
+                    raise TypeError("'function' in tool must be a dictionary")
+
+                for key in ["name", "description", "parameters"]:
+                    if not getattr(tool_obj.function, key):
+                        raise ValueError(
+                            f"Tool is missing required '{key}' in 'function'"
+                        )
+
+        self.gleaning_check()
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.operations.resolve.ResolveOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + + + + + + +
+ Source code in docetl/operations/resolve.py +
 31
+ 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
+470
+471
+472
+473
+474
+475
+476
+477
+478
+479
+480
+481
+482
+483
+484
+485
+486
+487
+488
+489
+490
+491
+492
+493
+494
+495
+496
+497
+498
+499
+500
+501
+502
+503
+504
+505
+506
+507
+508
+509
+510
+511
+512
+513
+514
+515
+516
+517
+518
+519
+520
+521
+522
+523
+524
+525
+526
+527
+528
+529
+530
+531
+532
+533
+534
+535
+536
+537
+538
+539
+540
+541
+542
+543
+544
+545
+546
+547
+548
+549
+550
+551
+552
+553
+554
+555
+556
+557
+558
+559
+560
+561
+562
+563
+564
+565
+566
+567
+568
+569
+570
+571
+572
+573
+574
+575
+576
+577
+578
+579
+580
+581
+582
+583
+584
+585
+586
+587
+588
+589
+590
+591
+592
+593
+594
+595
+596
+597
+598
+599
+600
+601
+602
+603
+604
+605
+606
+607
+608
+609
+610
+611
+612
+613
+614
+615
+616
+617
+618
+619
+620
+621
+622
+623
+624
+625
+626
+627
+628
+629
+630
+631
+632
+633
+634
+635
+636
+637
+638
+639
+640
+641
+642
+643
+644
+645
+646
+647
+648
+649
+650
+651
+652
+653
+654
+655
+656
+657
+658
+659
+660
+661
+662
+663
+664
+665
+666
+667
+668
+669
+670
+671
+672
+673
+674
+675
+676
+677
+678
+679
+680
+681
+682
+683
+684
+685
+686
+687
+688
+689
+690
+691
+692
+693
+694
+695
+696
+697
class ResolveOperation(BaseOperation):
+    class schema(BaseOperation.schema):
+        type: str = "resolve"
+        comparison_prompt: str
+        resolution_prompt: str
+        output: Optional[Dict[str, Any]] = None
+        embedding_model: Optional[str] = None
+        resolution_model: Optional[str] = None
+        comparison_model: Optional[str] = None
+        blocking_keys: Optional[List[str]] = None
+        blocking_threshold: Optional[float] = None
+        blocking_conditions: Optional[List[str]] = None
+        input: Optional[Dict[str, Any]] = None
+        embedding_batch_size: Optional[int] = None
+        compare_batch_size: Optional[int] = None
+        limit_comparisons: Optional[int] = None
+        optimize: Optional[bool] = None
+        timeout: Optional[int] = None
+        litellm_completion_kwargs: Dict[str, Any] = Field(default_factory=dict)
+        enable_observability: bool = False
+
+    def compare_pair(
+        self,
+        comparison_prompt: str,
+        model: str,
+        item1: Dict,
+        item2: Dict,
+        blocking_keys: List[str] = [],
+        timeout_seconds: int = 120,
+        max_retries_per_timeout: int = 2,
+    ) -> Tuple[bool, float, str]:
+        """
+        Compares two items using an LLM model to determine if they match.
+
+        Args:
+            comparison_prompt (str): The prompt template for comparison.
+            model (str): The LLM model to use for comparison.
+            item1 (Dict): The first item to compare.
+            item2 (Dict): The second item to compare.
+
+        Returns:
+            Tuple[bool, float, str]: A tuple containing a boolean indicating whether the items match, the cost of the comparison, and the prompt.
+        """
+        if blocking_keys:
+            if all(
+                key in item1
+                and key in item2
+                and str(item1[key]).lower() == str(item2[key]).lower()
+                for key in blocking_keys
+            ):
+                return True, 0, ""
+
+        prompt_template = Template(comparison_prompt)
+        prompt = prompt_template.render(input1=item1, input2=item2)
+        response = self.runner.api.call_llm(
+            model,
+            "compare",
+            [{"role": "user", "content": prompt}],
+            {"is_match": "bool"},
+            timeout_seconds=timeout_seconds,
+            max_retries_per_timeout=max_retries_per_timeout,
+            bypass_cache=self.config.get("bypass_cache", False),
+            litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+        )
+        output = self.runner.api.parse_llm_response(
+            response.response,
+            {"is_match": "bool"},
+        )[0]
+
+        return output["is_match"], response.total_cost, prompt
+
+    def syntax_check(self) -> None:
+        """
+        Checks the configuration of the ResolveOperation for required keys and valid structure.
+
+        This method performs the following checks:
+        1. Verifies the presence of required keys: 'comparison_prompt' and 'output'.
+        2. Ensures 'output' contains a 'schema' key.
+        3. Validates that 'schema' in 'output' is a non-empty dictionary.
+        4. Checks if 'comparison_prompt' is a valid Jinja2 template with 'input1' and 'input2' variables.
+        5. If 'resolution_prompt' is present, verifies it as a valid Jinja2 template with 'inputs' variable.
+        6. Optionally checks if 'model' is a string (if present).
+        7. Optionally checks 'blocking_keys' (if present, further checks are performed).
+
+        Raises:
+            ValueError: If required keys are missing, if templates are invalid or missing required variables,
+                        or if any other configuration aspect is incorrect or inconsistent.
+            TypeError: If the types of configuration values are incorrect, such as 'schema' not being a dict
+                       or 'model' not being a string.
+        """
+        required_keys = ["comparison_prompt", "output"]
+        for key in required_keys:
+            if key not in self.config:
+                raise ValueError(
+                    f"Missing required key '{key}' in ResolveOperation configuration"
+                )
+
+        if "schema" not in self.config["output"]:
+            raise ValueError("Missing 'schema' in 'output' configuration")
+
+        if not isinstance(self.config["output"]["schema"], dict):
+            raise TypeError("'schema' in 'output' configuration must be a dictionary")
+
+        if not self.config["output"]["schema"]:
+            raise ValueError("'schema' in 'output' configuration cannot be empty")
+
+        # Check if the comparison_prompt is a valid Jinja2 template
+        try:
+            comparison_template = Template(self.config["comparison_prompt"])
+            comparison_vars = comparison_template.environment.parse(
+                self.config["comparison_prompt"]
+            ).find_all(jinja2.nodes.Name)
+            comparison_var_names = {var.name for var in comparison_vars}
+            if (
+                "input1" not in comparison_var_names
+                or "input2" not in comparison_var_names
+            ):
+                raise ValueError(
+                    "'comparison_prompt' must contain both 'input1' and 'input2' variables"
+                )
+
+            if "resolution_prompt" in self.config:
+                reduction_template = Template(self.config["resolution_prompt"])
+                reduction_vars = reduction_template.environment.parse(
+                    self.config["resolution_prompt"]
+                ).find_all(jinja2.nodes.Name)
+                reduction_var_names = {var.name for var in reduction_vars}
+                if "inputs" not in reduction_var_names:
+                    raise ValueError(
+                        "'resolution_prompt' must contain 'inputs' variable"
+                    )
+        except Exception as e:
+            raise ValueError(f"Invalid Jinja2 template: {str(e)}")
+
+        # Check if the model is specified (optional)
+        if "model" in self.config and not isinstance(self.config["model"], str):
+            raise TypeError("'model' in configuration must be a string")
+
+        # Check blocking_keys (optional)
+        if "blocking_keys" in self.config:
+            if not isinstance(self.config["blocking_keys"], list):
+                raise TypeError("'blocking_keys' must be a list")
+            if not all(isinstance(key, str) for key in self.config["blocking_keys"]):
+                raise TypeError("All items in 'blocking_keys' must be strings")
+
+        # Check blocking_threshold (optional)
+        if "blocking_threshold" in self.config:
+            if not isinstance(self.config["blocking_threshold"], (int, float)):
+                raise TypeError("'blocking_threshold' must be a number")
+            if not 0 <= self.config["blocking_threshold"] <= 1:
+                raise ValueError("'blocking_threshold' must be between 0 and 1")
+
+        # Check blocking_conditions (optional)
+        if "blocking_conditions" in self.config:
+            if not isinstance(self.config["blocking_conditions"], list):
+                raise TypeError("'blocking_conditions' must be a list")
+            if not all(
+                isinstance(cond, str) for cond in self.config["blocking_conditions"]
+            ):
+                raise TypeError("All items in 'blocking_conditions' must be strings")
+
+        # Check if input schema is provided and valid (optional)
+        if "input" in self.config:
+            if "schema" not in self.config["input"]:
+                raise ValueError("Missing 'schema' in 'input' configuration")
+            if not isinstance(self.config["input"]["schema"], dict):
+                raise TypeError(
+                    "'schema' in 'input' configuration must be a dictionary"
+                )
+
+        # Check limit_comparisons (optional)
+        if "limit_comparisons" in self.config:
+            if not isinstance(self.config["limit_comparisons"], int):
+                raise TypeError("'limit_comparisons' must be an integer")
+            if self.config["limit_comparisons"] <= 0:
+                raise ValueError("'limit_comparisons' must be a positive integer")
+
+    def validation_fn(self, response: Dict[str, Any]):
+        output = self.runner.api.parse_llm_response(
+            response,
+            schema=self.config["output"]["schema"],
+        )[0]
+        if self.runner.api.validate_output(self.config, output, self.console):
+            return output, True
+        return output, False
+
+    def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+        """
+        Executes the resolve operation on the provided dataset.
+
+        Args:
+            input_data (List[Dict]): The dataset to resolve.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the resolved results and the total cost of the operation.
+
+        This method performs the following steps:
+        1. Initial blocking based on specified conditions and/or embedding similarity
+        2. Pairwise comparison of potentially matching entries using LLM
+        3. Clustering of matched entries
+        4. Resolution of each cluster into a single entry (if applicable)
+        5. Result aggregation and validation
+
+        The method also calculates and logs statistics such as comparisons saved by blocking and self-join selectivity.
+        """
+        if len(input_data) == 0:
+            return [], 0
+
+        # Initialize observability data for all items at the start
+        if self.config.get("enable_observability", False):
+            observability_key = f"_observability_{self.config['name']}"
+            for item in input_data:
+                if observability_key not in item:
+                    item[observability_key] = {
+                        "comparison_prompts": [],
+                        "resolution_prompt": None
+                    }
+
+        blocking_keys = self.config.get("blocking_keys", [])
+        blocking_threshold = self.config.get("blocking_threshold")
+        blocking_conditions = self.config.get("blocking_conditions", [])
+        if self.status:
+            self.status.stop()
+
+        if not blocking_threshold and not blocking_conditions:
+            # Prompt the user for confirmation
+            if not Confirm.ask(
+                f"[yellow]Warning: No blocking keys or conditions specified. "
+                f"This may result in a large number of comparisons. "
+                f"We recommend specifying at least one blocking key or condition, or using the optimizer to automatically come up with these. "
+                f"Do you want to continue without blocking?[/yellow]",
+                console=self.runner.console,
+            ):
+                raise ValueError("Operation cancelled by user.")
+
+        input_schema = self.config.get("input", {}).get("schema", {})
+        if not blocking_keys:
+            # Set them to all keys in the input data
+            blocking_keys = list(input_data[0].keys())
+        limit_comparisons = self.config.get("limit_comparisons")
+        total_cost = 0
+
+        def is_match(item1: Dict[str, Any], item2: Dict[str, Any]) -> bool:
+            return any(
+                eval(condition, {"input1": item1, "input2": item2})
+                for condition in blocking_conditions
+            )
+
+        # Calculate embeddings if blocking_threshold is set
+        embeddings = None
+        if blocking_threshold is not None:
+            embedding_model = self.config.get(
+                "embedding_model", "text-embedding-3-small"
+            )
+
+            def get_embeddings_batch(
+                items: List[Dict[str, Any]]
+            ) -> List[Tuple[List[float], float]]:
+                texts = [
+                    " ".join(str(item[key]) for key in blocking_keys if key in item)
+                    for item in items
+                ]
+                response = self.runner.api.gen_embedding(
+                    model=embedding_model, input=texts
+                )
+                return [
+                    (data["embedding"], completion_cost(response))
+                    for data in response["data"]
+                ]
+
+            embeddings = []
+            costs = []
+            with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+                for i in range(
+                    0, len(input_data), self.config.get("embedding_batch_size", 1000)
+                ):
+                    batch = input_data[
+                        i : i + self.config.get("embedding_batch_size", 1000)
+                    ]
+                    batch_results = list(executor.map(get_embeddings_batch, [batch]))
+
+                    for result in batch_results:
+                        embeddings.extend([r[0] for r in result])
+                        costs.extend([r[1] for r in result])
+
+                total_cost += sum(costs)
+
+        # Generate all pairs to compare, ensuring no duplicate comparisons
+        def get_unique_comparison_pairs() -> Tuple[List[Tuple[int, int]], Dict[Tuple[str, ...], List[int]]]:
+            # Create a mapping of values to their indices
+            value_to_indices: Dict[Tuple[str, ...], List[int]] = {}
+            for i, item in enumerate(input_data):
+                # Create a hashable key from the blocking keys
+                key = tuple(str(item.get(k, "")) for k in blocking_keys)
+                if key not in value_to_indices:
+                    value_to_indices[key] = []
+                value_to_indices[key].append(i)
+
+            # Generate pairs for comparison, comparing each unique value combination only once
+            comparison_pairs = []
+            keys = list(value_to_indices.keys())
+
+            # First, handle comparisons between different values
+            for i in range(len(keys)):
+                for j in range(i + 1, len(keys)):
+                    # Only need one comparison between different values
+                    idx1 = value_to_indices[keys[i]][0]
+                    idx2 = value_to_indices[keys[j]][0]
+                    if idx1 < idx2:  # Maintain ordering to avoid duplicates
+                        comparison_pairs.append((idx1, idx2))
+
+            return comparison_pairs, value_to_indices
+
+        comparison_pairs, value_to_indices = get_unique_comparison_pairs()
+
+        # Filter pairs based on blocking conditions
+        def meets_blocking_conditions(pair: Tuple[int, int]) -> bool:
+            i, j = pair
+            return (
+                is_match(input_data[i], input_data[j]) if blocking_conditions else False
+            )
+
+        blocked_pairs = list(filter(meets_blocking_conditions, comparison_pairs)) if blocking_conditions else comparison_pairs
+
+        # Apply limit_comparisons to blocked pairs
+        if limit_comparisons is not None and len(blocked_pairs) > limit_comparisons:
+            self.console.log(
+                f"Randomly sampling {limit_comparisons} pairs out of {len(blocked_pairs)} blocked pairs."
+            )
+            blocked_pairs = random.sample(blocked_pairs, limit_comparisons)
+
+        # Initialize clusters with all indices
+        clusters = [{i} for i in range(len(input_data))]
+        cluster_map = {i: i for i in range(len(input_data))}
+
+        # If there are remaining comparisons, fill with highest cosine similarities
+        remaining_comparisons = (
+            limit_comparisons - len(blocked_pairs)
+            if limit_comparisons is not None
+            else float("inf")
+        )
+        if remaining_comparisons > 0 and blocking_threshold is not None:
+            # Compute cosine similarity for all pairs efficiently
+            from sklearn.metrics.pairwise import cosine_similarity
+
+            similarity_matrix = cosine_similarity(embeddings)
+
+            cosine_pairs = []
+            for i, j in comparison_pairs:
+                if (i, j) not in blocked_pairs and find_cluster(
+                    i, cluster_map
+                ) != find_cluster(j, cluster_map):
+                    similarity = similarity_matrix[i, j]
+                    if similarity >= blocking_threshold:
+                        cosine_pairs.append((i, j, similarity))
+
+            if remaining_comparisons != float("inf"):
+                cosine_pairs.sort(key=lambda x: x[2], reverse=True)
+                additional_pairs = [
+                    (i, j) for i, j, _ in cosine_pairs[: int(remaining_comparisons)]
+                ]
+                blocked_pairs.extend(additional_pairs)
+            else:
+                blocked_pairs.extend((i, j) for i, j, _ in cosine_pairs)
+
+        # Modified merge_clusters to handle all indices with the same value
+
+        def merge_clusters(item1: int, item2: int) -> None:
+            root1, root2 = find_cluster(item1, cluster_map), find_cluster(
+                item2, cluster_map
+            )
+            if root1 != root2:
+                if len(clusters[root1]) < len(clusters[root2]):
+                    root1, root2 = root2, root1
+                clusters[root1] |= clusters[root2]
+                cluster_map[root2] = root1
+                clusters[root2] = set()
+
+                # Also merge all other indices that share the same values
+                key1 = tuple(str(input_data[item1].get(k, "")) for k in blocking_keys)
+                key2 = tuple(str(input_data[item2].get(k, "")) for k in blocking_keys)
+
+                # Merge all indices with the same values
+                for idx in value_to_indices.get(key1, []):
+                    if idx != item1:
+                        root_idx = find_cluster(idx, cluster_map)
+                        if root_idx != root1:
+                            clusters[root1] |= clusters[root_idx]
+                            cluster_map[root_idx] = root1
+                            clusters[root_idx] = set()
+
+                for idx in value_to_indices.get(key2, []):
+                    if idx != item2:
+                        root_idx = find_cluster(idx, cluster_map)
+                        if root_idx != root1:
+                            clusters[root1] |= clusters[root_idx]
+                            cluster_map[root_idx] = root1
+                            clusters[root_idx] = set()
+
+        # Calculate and print statistics
+        total_possible_comparisons = len(input_data) * (len(input_data) - 1) // 2
+        comparisons_made = len(blocked_pairs)
+        comparisons_saved = total_possible_comparisons - comparisons_made
+        self.console.log(
+            f"[green]Comparisons saved by blocking: {comparisons_saved} "
+            f"({(comparisons_saved / total_possible_comparisons) * 100:.2f}%)[/green]"
+        )
+        self.console.log(
+            f"[blue]Number of pairs to compare: {len(blocked_pairs)}[/blue]"
+        )
+
+        # Compute an auto-batch size based on the number of comparisons
+        def auto_batch() -> int:
+            # Maximum batch size limit for 4o-mini model
+            M = 500
+
+            n = len(input_data)
+            m = len(blocked_pairs)
+
+            # https://www.wolframalpha.com/input?i=k%28k-1%29%2F2+%2B+%28n-k%29%28k-1%29+%3D+m%2C+solve+for+k
+            # Two possible solutions for k:
+            # k = -1/2 sqrt((1 - 2n)^2 - 8m) + n + 1/2
+            # k = 1/2 (sqrt((1 - 2n)^2 - 8m) + 2n + 1)
+
+            discriminant = (1 - 2*n)**2 - 8*m
+            sqrt_discriminant = discriminant ** 0.5
+
+            k1 = -0.5 * sqrt_discriminant + n + 0.5
+            k2 = 0.5 * (sqrt_discriminant + 2*n + 1)
+
+            # Take the maximum viable solution
+            k = max(k1, k2)
+            return M if k < 0 else min(int(k), M)
+
+        # Compare pairs and update clusters in real-time
+        batch_size = self.config.get("compare_batch_size", auto_batch())
+        self.console.log(f"Using compare batch size: {batch_size}")
+        pair_costs = 0
+
+        pbar = RichLoopBar(
+            range(0, len(blocked_pairs), batch_size),
+            desc=f"Processing batches of {batch_size} LLM comparisons",
+            console=self.console,
+        )
+        last_processed = 0
+        for i in pbar:
+            batch_end = last_processed + batch_size
+            batch = blocked_pairs[last_processed : batch_end]
+            # Filter pairs for the initial batch
+            better_batch = [
+                pair for pair in batch
+                if find_cluster(pair[0], cluster_map) == pair[0] and find_cluster(pair[1], cluster_map) == pair[1]
+            ]
+
+            # Expand better_batch if it doesn’t reach batch_size
+            while len(better_batch) < batch_size and batch_end < len(blocked_pairs):
+                # Move batch_end forward by batch_size to get more pairs
+                next_end = batch_end + batch_size
+                next_batch = blocked_pairs[batch_end:next_end]
+
+                better_batch.extend(
+                    pair for pair in next_batch
+                    if find_cluster(pair[0], cluster_map) == pair[0] and find_cluster(pair[1], cluster_map) == pair[1]
+                )
+
+                # Update batch_end to prevent overlapping in the next loop
+                batch_end = next_end
+            better_batch = better_batch[:batch_size]
+            last_processed = batch_end
+            with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+                future_to_pair = {
+                    executor.submit(
+                        self.compare_pair,
+                        self.config["comparison_prompt"],
+                        self.config.get("comparison_model", self.default_model),
+                        input_data[pair[0]],
+                        input_data[pair[1]],
+                        blocking_keys,
+                        timeout_seconds=self.config.get("timeout", 120),
+                        max_retries_per_timeout=self.config.get(
+                            "max_retries_per_timeout", 2
+                        ),
+                    ): pair
+                    for pair in better_batch
+                }
+
+                for future in as_completed(future_to_pair):
+                    pair = future_to_pair[future]
+                    is_match_result, cost, prompt = future.result()
+                    pair_costs += cost
+                    if is_match_result:
+                        merge_clusters(pair[0], pair[1])
+
+                    if self.config.get("enable_observability", False):
+                        observability_key = f"_observability_{self.config['name']}"
+                        for idx in (pair[0], pair[1]):
+                            if observability_key not in input_data[idx]:
+                                input_data[idx][observability_key] = {
+                                    "comparison_prompts": [],
+                                    "resolution_prompt": None
+                                }
+                            input_data[idx][observability_key]["comparison_prompts"].append(prompt)
+
+                    pbar.update(last_processed//batch_size)
+        total_cost += pair_costs
+
+        # Collect final clusters
+        final_clusters = [cluster for cluster in clusters if cluster]
+
+        # Process each cluster
+        results = []
+
+        def process_cluster(cluster):
+            if len(cluster) > 1:
+                cluster_items = [input_data[i] for i in cluster]
+                reduction_template = Template(self.config["resolution_prompt"])
+                if input_schema:
+                    cluster_items = [
+                        {k: item[k] for k in input_schema.keys() if k in item}
+                        for item in cluster_items
+                    ]
+
+                resolution_prompt = reduction_template.render(inputs=cluster_items)
+                reduction_response = self.runner.api.call_llm(
+                    self.config.get("resolution_model", self.default_model),
+                    "reduce",
+                    [{"role": "user", "content": resolution_prompt}],
+                    self.config["output"]["schema"],
+                    timeout_seconds=self.config.get("timeout", 120),
+                    max_retries_per_timeout=self.config.get(
+                        "max_retries_per_timeout", 2
+                    ),
+                    bypass_cache=self.config.get("bypass_cache", False),
+                    validation_config=(
+                        {
+                            "val_rule": self.config.get("validate", []),
+                            "validation_fn": self.validation_fn,
+                        }
+                        if self.config.get("validate", None)
+                        else None
+                    ),
+                    litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+                )
+                reduction_cost = reduction_response.total_cost
+
+                if self.config.get("enable_observability", False):
+                    for item in [input_data[i] for i in cluster]:
+                        observability_key = f"_observability_{self.config['name']}"
+                        if observability_key not in item:
+                            item[observability_key] = {
+                                "comparison_prompts": [],
+                                "resolution_prompt": None
+                            }
+                        item[observability_key]["resolution_prompt"] = resolution_prompt
+
+                if reduction_response.validated:
+                    reduction_output = self.runner.api.parse_llm_response(
+                        reduction_response.response,
+                        self.config["output"]["schema"],
+                        manually_fix_errors=self.manually_fix_errors,
+                    )[0]
+
+                    # If the output is overwriting an existing key, we want to save the kv pairs
+                    keys_in_output = [
+                        k
+                        for k in set(reduction_output.keys())
+                        if k in cluster_items[0].keys()
+                    ]
+
+                    return (
+                        [
+                            {
+                                **item,
+                                f"_kv_pairs_preresolve_{self.config['name']}": {
+                                    k: item[k] for k in keys_in_output
+                                },
+                                **{
+                                    k: reduction_output[k]
+                                    for k in self.config["output"]["schema"]
+                                },
+                            }
+                            for item in [input_data[i] for i in cluster]
+                        ],
+                        reduction_cost,
+                    )
+                return [], reduction_cost
+            else:
+                # Set the output schema to be the keys found in the compare_prompt
+                compare_prompt_keys = extract_jinja_variables(
+                    self.config["comparison_prompt"]
+                )
+                # Get the set of keys in the compare_prompt
+                compare_prompt_keys = set(
+                    [
+                        k.replace("input1.", "")
+                        for k in compare_prompt_keys
+                        if "input1" in k
+                    ]
+                )
+
+                # For each key in the output schema, find the most similar key in the compare_prompt
+                output_keys = set(self.config["output"]["schema"].keys())
+                key_mapping = {}
+                for output_key in output_keys:
+                    best_match = None
+                    best_score = 0
+                    for compare_key in compare_prompt_keys:
+                        score = sum(
+                            c1 == c2 for c1, c2 in zip(output_key, compare_key)
+                        ) / max(len(output_key), len(compare_key))
+                        if score > best_score:
+                            best_score = score
+                            best_match = compare_key
+                    key_mapping[output_key] = best_match
+
+                # Create the result dictionary using the key mapping
+                result = input_data[list(cluster)[0]].copy()
+                result[f"_kv_pairs_preresolve_{self.config['name']}"] = {
+                    ok: result[ck] for ok, ck in key_mapping.items() if ck in result
+                }
+                for output_key, compare_key in key_mapping.items():
+                    if compare_key in input_data[list(cluster)[0]]:
+                        result[output_key] = input_data[list(cluster)[0]][compare_key]
+                    elif output_key in input_data[list(cluster)[0]]:
+                        result[output_key] = input_data[list(cluster)[0]][output_key]
+                    else:
+                        result[output_key] = None  # or some default value
+
+                return [result], 0
+
+        # Calculate the number of records before and clusters after
+        num_records_before = len(input_data)
+        num_clusters_after = len(final_clusters)
+        self.console.log(f"Number of keys before resolution: {num_records_before}")
+        self.console.log(
+            f"Number of distinct keys after resolution: {num_clusters_after}"
+        )
+
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            futures = [
+                executor.submit(process_cluster, cluster) for cluster in final_clusters
+            ]
+            for future in rich_as_completed(
+                futures,
+                total=len(futures),
+                desc="Determining resolved key for each group of equivalent keys",
+                console=self.console,
+            ):
+                cluster_results, cluster_cost = future.result()
+                results.extend(cluster_results)
+                total_cost += cluster_cost
+
+        total_pairs = len(input_data) * (len(input_data) - 1) // 2
+        true_match_count = sum(
+            len(cluster) * (len(cluster) - 1) // 2
+            for cluster in final_clusters
+            if len(cluster) > 1
+        )
+        true_match_selectivity = (
+            true_match_count / total_pairs if total_pairs > 0 else 0
+        )
+        self.console.log(f"Self-join selectivity: {true_match_selectivity:.4f}")
+
+        if self.status:
+            self.status.start()
+
+        return results, total_cost
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ compare_pair(comparison_prompt, model, item1, item2, blocking_keys=[], timeout_seconds=120, max_retries_per_timeout=2) + +

+ + +
+ +

Compares two items using an LLM model to determine if they match.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ comparison_prompt + + str + +
+

The prompt template for comparison.

+
+
+ required +
+ model + + str + +
+

The LLM model to use for comparison.

+
+
+ required +
+ item1 + + Dict + +
+

The first item to compare.

+
+
+ required +
+ item2 + + Dict + +
+

The second item to compare.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[bool, float, str] + +
+

Tuple[bool, float, str]: A tuple containing a boolean indicating whether the items match, the cost of the comparison, and the prompt.

+
+
+ +
+ Source code in docetl/operations/resolve.py +
 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
def compare_pair(
+    self,
+    comparison_prompt: str,
+    model: str,
+    item1: Dict,
+    item2: Dict,
+    blocking_keys: List[str] = [],
+    timeout_seconds: int = 120,
+    max_retries_per_timeout: int = 2,
+) -> Tuple[bool, float, str]:
+    """
+    Compares two items using an LLM model to determine if they match.
+
+    Args:
+        comparison_prompt (str): The prompt template for comparison.
+        model (str): The LLM model to use for comparison.
+        item1 (Dict): The first item to compare.
+        item2 (Dict): The second item to compare.
+
+    Returns:
+        Tuple[bool, float, str]: A tuple containing a boolean indicating whether the items match, the cost of the comparison, and the prompt.
+    """
+    if blocking_keys:
+        if all(
+            key in item1
+            and key in item2
+            and str(item1[key]).lower() == str(item2[key]).lower()
+            for key in blocking_keys
+        ):
+            return True, 0, ""
+
+    prompt_template = Template(comparison_prompt)
+    prompt = prompt_template.render(input1=item1, input2=item2)
+    response = self.runner.api.call_llm(
+        model,
+        "compare",
+        [{"role": "user", "content": prompt}],
+        {"is_match": "bool"},
+        timeout_seconds=timeout_seconds,
+        max_retries_per_timeout=max_retries_per_timeout,
+        bypass_cache=self.config.get("bypass_cache", False),
+        litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+    )
+    output = self.runner.api.parse_llm_response(
+        response.response,
+        {"is_match": "bool"},
+    )[0]
+
+    return output["is_match"], response.total_cost, prompt
+
+
+
+ +
+ +
+ + +

+ execute(input_data) + +

+ + +
+ +

Executes the resolve operation on the provided dataset.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ input_data + + List[Dict] + +
+

The dataset to resolve.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[List[Dict], float] + +
+

Tuple[List[Dict], float]: A tuple containing the resolved results and the total cost of the operation.

+
+
+

This method performs the following steps: +1. Initial blocking based on specified conditions and/or embedding similarity +2. Pairwise comparison of potentially matching entries using LLM +3. Clustering of matched entries +4. Resolution of each cluster into a single entry (if applicable) +5. Result aggregation and validation

+

The method also calculates and logs statistics such as comparisons saved by blocking and self-join selectivity.

+ +
+ Source code in docetl/operations/resolve.py +
217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
+470
+471
+472
+473
+474
+475
+476
+477
+478
+479
+480
+481
+482
+483
+484
+485
+486
+487
+488
+489
+490
+491
+492
+493
+494
+495
+496
+497
+498
+499
+500
+501
+502
+503
+504
+505
+506
+507
+508
+509
+510
+511
+512
+513
+514
+515
+516
+517
+518
+519
+520
+521
+522
+523
+524
+525
+526
+527
+528
+529
+530
+531
+532
+533
+534
+535
+536
+537
+538
+539
+540
+541
+542
+543
+544
+545
+546
+547
+548
+549
+550
+551
+552
+553
+554
+555
+556
+557
+558
+559
+560
+561
+562
+563
+564
+565
+566
+567
+568
+569
+570
+571
+572
+573
+574
+575
+576
+577
+578
+579
+580
+581
+582
+583
+584
+585
+586
+587
+588
+589
+590
+591
+592
+593
+594
+595
+596
+597
+598
+599
+600
+601
+602
+603
+604
+605
+606
+607
+608
+609
+610
+611
+612
+613
+614
+615
+616
+617
+618
+619
+620
+621
+622
+623
+624
+625
+626
+627
+628
+629
+630
+631
+632
+633
+634
+635
+636
+637
+638
+639
+640
+641
+642
+643
+644
+645
+646
+647
+648
+649
+650
+651
+652
+653
+654
+655
+656
+657
+658
+659
+660
+661
+662
+663
+664
+665
+666
+667
+668
+669
+670
+671
+672
+673
+674
+675
+676
+677
+678
+679
+680
+681
+682
+683
+684
+685
+686
+687
+688
+689
+690
+691
+692
+693
+694
+695
+696
+697
def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+    """
+    Executes the resolve operation on the provided dataset.
+
+    Args:
+        input_data (List[Dict]): The dataset to resolve.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the resolved results and the total cost of the operation.
+
+    This method performs the following steps:
+    1. Initial blocking based on specified conditions and/or embedding similarity
+    2. Pairwise comparison of potentially matching entries using LLM
+    3. Clustering of matched entries
+    4. Resolution of each cluster into a single entry (if applicable)
+    5. Result aggregation and validation
+
+    The method also calculates and logs statistics such as comparisons saved by blocking and self-join selectivity.
+    """
+    if len(input_data) == 0:
+        return [], 0
+
+    # Initialize observability data for all items at the start
+    if self.config.get("enable_observability", False):
+        observability_key = f"_observability_{self.config['name']}"
+        for item in input_data:
+            if observability_key not in item:
+                item[observability_key] = {
+                    "comparison_prompts": [],
+                    "resolution_prompt": None
+                }
+
+    blocking_keys = self.config.get("blocking_keys", [])
+    blocking_threshold = self.config.get("blocking_threshold")
+    blocking_conditions = self.config.get("blocking_conditions", [])
+    if self.status:
+        self.status.stop()
+
+    if not blocking_threshold and not blocking_conditions:
+        # Prompt the user for confirmation
+        if not Confirm.ask(
+            f"[yellow]Warning: No blocking keys or conditions specified. "
+            f"This may result in a large number of comparisons. "
+            f"We recommend specifying at least one blocking key or condition, or using the optimizer to automatically come up with these. "
+            f"Do you want to continue without blocking?[/yellow]",
+            console=self.runner.console,
+        ):
+            raise ValueError("Operation cancelled by user.")
+
+    input_schema = self.config.get("input", {}).get("schema", {})
+    if not blocking_keys:
+        # Set them to all keys in the input data
+        blocking_keys = list(input_data[0].keys())
+    limit_comparisons = self.config.get("limit_comparisons")
+    total_cost = 0
+
+    def is_match(item1: Dict[str, Any], item2: Dict[str, Any]) -> bool:
+        return any(
+            eval(condition, {"input1": item1, "input2": item2})
+            for condition in blocking_conditions
+        )
+
+    # Calculate embeddings if blocking_threshold is set
+    embeddings = None
+    if blocking_threshold is not None:
+        embedding_model = self.config.get(
+            "embedding_model", "text-embedding-3-small"
+        )
+
+        def get_embeddings_batch(
+            items: List[Dict[str, Any]]
+        ) -> List[Tuple[List[float], float]]:
+            texts = [
+                " ".join(str(item[key]) for key in blocking_keys if key in item)
+                for item in items
+            ]
+            response = self.runner.api.gen_embedding(
+                model=embedding_model, input=texts
+            )
+            return [
+                (data["embedding"], completion_cost(response))
+                for data in response["data"]
+            ]
+
+        embeddings = []
+        costs = []
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            for i in range(
+                0, len(input_data), self.config.get("embedding_batch_size", 1000)
+            ):
+                batch = input_data[
+                    i : i + self.config.get("embedding_batch_size", 1000)
+                ]
+                batch_results = list(executor.map(get_embeddings_batch, [batch]))
+
+                for result in batch_results:
+                    embeddings.extend([r[0] for r in result])
+                    costs.extend([r[1] for r in result])
+
+            total_cost += sum(costs)
+
+    # Generate all pairs to compare, ensuring no duplicate comparisons
+    def get_unique_comparison_pairs() -> Tuple[List[Tuple[int, int]], Dict[Tuple[str, ...], List[int]]]:
+        # Create a mapping of values to their indices
+        value_to_indices: Dict[Tuple[str, ...], List[int]] = {}
+        for i, item in enumerate(input_data):
+            # Create a hashable key from the blocking keys
+            key = tuple(str(item.get(k, "")) for k in blocking_keys)
+            if key not in value_to_indices:
+                value_to_indices[key] = []
+            value_to_indices[key].append(i)
+
+        # Generate pairs for comparison, comparing each unique value combination only once
+        comparison_pairs = []
+        keys = list(value_to_indices.keys())
+
+        # First, handle comparisons between different values
+        for i in range(len(keys)):
+            for j in range(i + 1, len(keys)):
+                # Only need one comparison between different values
+                idx1 = value_to_indices[keys[i]][0]
+                idx2 = value_to_indices[keys[j]][0]
+                if idx1 < idx2:  # Maintain ordering to avoid duplicates
+                    comparison_pairs.append((idx1, idx2))
+
+        return comparison_pairs, value_to_indices
+
+    comparison_pairs, value_to_indices = get_unique_comparison_pairs()
+
+    # Filter pairs based on blocking conditions
+    def meets_blocking_conditions(pair: Tuple[int, int]) -> bool:
+        i, j = pair
+        return (
+            is_match(input_data[i], input_data[j]) if blocking_conditions else False
+        )
+
+    blocked_pairs = list(filter(meets_blocking_conditions, comparison_pairs)) if blocking_conditions else comparison_pairs
+
+    # Apply limit_comparisons to blocked pairs
+    if limit_comparisons is not None and len(blocked_pairs) > limit_comparisons:
+        self.console.log(
+            f"Randomly sampling {limit_comparisons} pairs out of {len(blocked_pairs)} blocked pairs."
+        )
+        blocked_pairs = random.sample(blocked_pairs, limit_comparisons)
+
+    # Initialize clusters with all indices
+    clusters = [{i} for i in range(len(input_data))]
+    cluster_map = {i: i for i in range(len(input_data))}
+
+    # If there are remaining comparisons, fill with highest cosine similarities
+    remaining_comparisons = (
+        limit_comparisons - len(blocked_pairs)
+        if limit_comparisons is not None
+        else float("inf")
+    )
+    if remaining_comparisons > 0 and blocking_threshold is not None:
+        # Compute cosine similarity for all pairs efficiently
+        from sklearn.metrics.pairwise import cosine_similarity
+
+        similarity_matrix = cosine_similarity(embeddings)
+
+        cosine_pairs = []
+        for i, j in comparison_pairs:
+            if (i, j) not in blocked_pairs and find_cluster(
+                i, cluster_map
+            ) != find_cluster(j, cluster_map):
+                similarity = similarity_matrix[i, j]
+                if similarity >= blocking_threshold:
+                    cosine_pairs.append((i, j, similarity))
+
+        if remaining_comparisons != float("inf"):
+            cosine_pairs.sort(key=lambda x: x[2], reverse=True)
+            additional_pairs = [
+                (i, j) for i, j, _ in cosine_pairs[: int(remaining_comparisons)]
+            ]
+            blocked_pairs.extend(additional_pairs)
+        else:
+            blocked_pairs.extend((i, j) for i, j, _ in cosine_pairs)
+
+    # Modified merge_clusters to handle all indices with the same value
+
+    def merge_clusters(item1: int, item2: int) -> None:
+        root1, root2 = find_cluster(item1, cluster_map), find_cluster(
+            item2, cluster_map
+        )
+        if root1 != root2:
+            if len(clusters[root1]) < len(clusters[root2]):
+                root1, root2 = root2, root1
+            clusters[root1] |= clusters[root2]
+            cluster_map[root2] = root1
+            clusters[root2] = set()
+
+            # Also merge all other indices that share the same values
+            key1 = tuple(str(input_data[item1].get(k, "")) for k in blocking_keys)
+            key2 = tuple(str(input_data[item2].get(k, "")) for k in blocking_keys)
+
+            # Merge all indices with the same values
+            for idx in value_to_indices.get(key1, []):
+                if idx != item1:
+                    root_idx = find_cluster(idx, cluster_map)
+                    if root_idx != root1:
+                        clusters[root1] |= clusters[root_idx]
+                        cluster_map[root_idx] = root1
+                        clusters[root_idx] = set()
+
+            for idx in value_to_indices.get(key2, []):
+                if idx != item2:
+                    root_idx = find_cluster(idx, cluster_map)
+                    if root_idx != root1:
+                        clusters[root1] |= clusters[root_idx]
+                        cluster_map[root_idx] = root1
+                        clusters[root_idx] = set()
+
+    # Calculate and print statistics
+    total_possible_comparisons = len(input_data) * (len(input_data) - 1) // 2
+    comparisons_made = len(blocked_pairs)
+    comparisons_saved = total_possible_comparisons - comparisons_made
+    self.console.log(
+        f"[green]Comparisons saved by blocking: {comparisons_saved} "
+        f"({(comparisons_saved / total_possible_comparisons) * 100:.2f}%)[/green]"
+    )
+    self.console.log(
+        f"[blue]Number of pairs to compare: {len(blocked_pairs)}[/blue]"
+    )
+
+    # Compute an auto-batch size based on the number of comparisons
+    def auto_batch() -> int:
+        # Maximum batch size limit for 4o-mini model
+        M = 500
+
+        n = len(input_data)
+        m = len(blocked_pairs)
+
+        # https://www.wolframalpha.com/input?i=k%28k-1%29%2F2+%2B+%28n-k%29%28k-1%29+%3D+m%2C+solve+for+k
+        # Two possible solutions for k:
+        # k = -1/2 sqrt((1 - 2n)^2 - 8m) + n + 1/2
+        # k = 1/2 (sqrt((1 - 2n)^2 - 8m) + 2n + 1)
+
+        discriminant = (1 - 2*n)**2 - 8*m
+        sqrt_discriminant = discriminant ** 0.5
+
+        k1 = -0.5 * sqrt_discriminant + n + 0.5
+        k2 = 0.5 * (sqrt_discriminant + 2*n + 1)
+
+        # Take the maximum viable solution
+        k = max(k1, k2)
+        return M if k < 0 else min(int(k), M)
+
+    # Compare pairs and update clusters in real-time
+    batch_size = self.config.get("compare_batch_size", auto_batch())
+    self.console.log(f"Using compare batch size: {batch_size}")
+    pair_costs = 0
+
+    pbar = RichLoopBar(
+        range(0, len(blocked_pairs), batch_size),
+        desc=f"Processing batches of {batch_size} LLM comparisons",
+        console=self.console,
+    )
+    last_processed = 0
+    for i in pbar:
+        batch_end = last_processed + batch_size
+        batch = blocked_pairs[last_processed : batch_end]
+        # Filter pairs for the initial batch
+        better_batch = [
+            pair for pair in batch
+            if find_cluster(pair[0], cluster_map) == pair[0] and find_cluster(pair[1], cluster_map) == pair[1]
+        ]
+
+        # Expand better_batch if it doesn’t reach batch_size
+        while len(better_batch) < batch_size and batch_end < len(blocked_pairs):
+            # Move batch_end forward by batch_size to get more pairs
+            next_end = batch_end + batch_size
+            next_batch = blocked_pairs[batch_end:next_end]
+
+            better_batch.extend(
+                pair for pair in next_batch
+                if find_cluster(pair[0], cluster_map) == pair[0] and find_cluster(pair[1], cluster_map) == pair[1]
+            )
+
+            # Update batch_end to prevent overlapping in the next loop
+            batch_end = next_end
+        better_batch = better_batch[:batch_size]
+        last_processed = batch_end
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            future_to_pair = {
+                executor.submit(
+                    self.compare_pair,
+                    self.config["comparison_prompt"],
+                    self.config.get("comparison_model", self.default_model),
+                    input_data[pair[0]],
+                    input_data[pair[1]],
+                    blocking_keys,
+                    timeout_seconds=self.config.get("timeout", 120),
+                    max_retries_per_timeout=self.config.get(
+                        "max_retries_per_timeout", 2
+                    ),
+                ): pair
+                for pair in better_batch
+            }
+
+            for future in as_completed(future_to_pair):
+                pair = future_to_pair[future]
+                is_match_result, cost, prompt = future.result()
+                pair_costs += cost
+                if is_match_result:
+                    merge_clusters(pair[0], pair[1])
+
+                if self.config.get("enable_observability", False):
+                    observability_key = f"_observability_{self.config['name']}"
+                    for idx in (pair[0], pair[1]):
+                        if observability_key not in input_data[idx]:
+                            input_data[idx][observability_key] = {
+                                "comparison_prompts": [],
+                                "resolution_prompt": None
+                            }
+                        input_data[idx][observability_key]["comparison_prompts"].append(prompt)
+
+                pbar.update(last_processed//batch_size)
+    total_cost += pair_costs
+
+    # Collect final clusters
+    final_clusters = [cluster for cluster in clusters if cluster]
+
+    # Process each cluster
+    results = []
+
+    def process_cluster(cluster):
+        if len(cluster) > 1:
+            cluster_items = [input_data[i] for i in cluster]
+            reduction_template = Template(self.config["resolution_prompt"])
+            if input_schema:
+                cluster_items = [
+                    {k: item[k] for k in input_schema.keys() if k in item}
+                    for item in cluster_items
+                ]
+
+            resolution_prompt = reduction_template.render(inputs=cluster_items)
+            reduction_response = self.runner.api.call_llm(
+                self.config.get("resolution_model", self.default_model),
+                "reduce",
+                [{"role": "user", "content": resolution_prompt}],
+                self.config["output"]["schema"],
+                timeout_seconds=self.config.get("timeout", 120),
+                max_retries_per_timeout=self.config.get(
+                    "max_retries_per_timeout", 2
+                ),
+                bypass_cache=self.config.get("bypass_cache", False),
+                validation_config=(
+                    {
+                        "val_rule": self.config.get("validate", []),
+                        "validation_fn": self.validation_fn,
+                    }
+                    if self.config.get("validate", None)
+                    else None
+                ),
+                litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+            )
+            reduction_cost = reduction_response.total_cost
+
+            if self.config.get("enable_observability", False):
+                for item in [input_data[i] for i in cluster]:
+                    observability_key = f"_observability_{self.config['name']}"
+                    if observability_key not in item:
+                        item[observability_key] = {
+                            "comparison_prompts": [],
+                            "resolution_prompt": None
+                        }
+                    item[observability_key]["resolution_prompt"] = resolution_prompt
+
+            if reduction_response.validated:
+                reduction_output = self.runner.api.parse_llm_response(
+                    reduction_response.response,
+                    self.config["output"]["schema"],
+                    manually_fix_errors=self.manually_fix_errors,
+                )[0]
+
+                # If the output is overwriting an existing key, we want to save the kv pairs
+                keys_in_output = [
+                    k
+                    for k in set(reduction_output.keys())
+                    if k in cluster_items[0].keys()
+                ]
+
+                return (
+                    [
+                        {
+                            **item,
+                            f"_kv_pairs_preresolve_{self.config['name']}": {
+                                k: item[k] for k in keys_in_output
+                            },
+                            **{
+                                k: reduction_output[k]
+                                for k in self.config["output"]["schema"]
+                            },
+                        }
+                        for item in [input_data[i] for i in cluster]
+                    ],
+                    reduction_cost,
+                )
+            return [], reduction_cost
+        else:
+            # Set the output schema to be the keys found in the compare_prompt
+            compare_prompt_keys = extract_jinja_variables(
+                self.config["comparison_prompt"]
+            )
+            # Get the set of keys in the compare_prompt
+            compare_prompt_keys = set(
+                [
+                    k.replace("input1.", "")
+                    for k in compare_prompt_keys
+                    if "input1" in k
+                ]
+            )
+
+            # For each key in the output schema, find the most similar key in the compare_prompt
+            output_keys = set(self.config["output"]["schema"].keys())
+            key_mapping = {}
+            for output_key in output_keys:
+                best_match = None
+                best_score = 0
+                for compare_key in compare_prompt_keys:
+                    score = sum(
+                        c1 == c2 for c1, c2 in zip(output_key, compare_key)
+                    ) / max(len(output_key), len(compare_key))
+                    if score > best_score:
+                        best_score = score
+                        best_match = compare_key
+                key_mapping[output_key] = best_match
+
+            # Create the result dictionary using the key mapping
+            result = input_data[list(cluster)[0]].copy()
+            result[f"_kv_pairs_preresolve_{self.config['name']}"] = {
+                ok: result[ck] for ok, ck in key_mapping.items() if ck in result
+            }
+            for output_key, compare_key in key_mapping.items():
+                if compare_key in input_data[list(cluster)[0]]:
+                    result[output_key] = input_data[list(cluster)[0]][compare_key]
+                elif output_key in input_data[list(cluster)[0]]:
+                    result[output_key] = input_data[list(cluster)[0]][output_key]
+                else:
+                    result[output_key] = None  # or some default value
+
+            return [result], 0
+
+    # Calculate the number of records before and clusters after
+    num_records_before = len(input_data)
+    num_clusters_after = len(final_clusters)
+    self.console.log(f"Number of keys before resolution: {num_records_before}")
+    self.console.log(
+        f"Number of distinct keys after resolution: {num_clusters_after}"
+    )
+
+    with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+        futures = [
+            executor.submit(process_cluster, cluster) for cluster in final_clusters
+        ]
+        for future in rich_as_completed(
+            futures,
+            total=len(futures),
+            desc="Determining resolved key for each group of equivalent keys",
+            console=self.console,
+        ):
+            cluster_results, cluster_cost = future.result()
+            results.extend(cluster_results)
+            total_cost += cluster_cost
+
+    total_pairs = len(input_data) * (len(input_data) - 1) // 2
+    true_match_count = sum(
+        len(cluster) * (len(cluster) - 1) // 2
+        for cluster in final_clusters
+        if len(cluster) > 1
+    )
+    true_match_selectivity = (
+        true_match_count / total_pairs if total_pairs > 0 else 0
+    )
+    self.console.log(f"Self-join selectivity: {true_match_selectivity:.4f}")
+
+    if self.status:
+        self.status.start()
+
+    return results, total_cost
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Checks the configuration of the ResolveOperation for required keys and valid structure.

+

This method performs the following checks: +1. Verifies the presence of required keys: 'comparison_prompt' and 'output'. +2. Ensures 'output' contains a 'schema' key. +3. Validates that 'schema' in 'output' is a non-empty dictionary. +4. Checks if 'comparison_prompt' is a valid Jinja2 template with 'input1' and 'input2' variables. +5. If 'resolution_prompt' is present, verifies it as a valid Jinja2 template with 'inputs' variable. +6. Optionally checks if 'model' is a string (if present). +7. Optionally checks 'blocking_keys' (if present, further checks are performed).

+ + +

Raises:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If required keys are missing, if templates are invalid or missing required variables, + or if any other configuration aspect is incorrect or inconsistent.

+
+
+ TypeError + +
+

If the types of configuration values are incorrect, such as 'schema' not being a dict + or 'model' not being a string.

+
+
+ +
+ Source code in docetl/operations/resolve.py +
102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
def syntax_check(self) -> None:
+    """
+    Checks the configuration of the ResolveOperation for required keys and valid structure.
+
+    This method performs the following checks:
+    1. Verifies the presence of required keys: 'comparison_prompt' and 'output'.
+    2. Ensures 'output' contains a 'schema' key.
+    3. Validates that 'schema' in 'output' is a non-empty dictionary.
+    4. Checks if 'comparison_prompt' is a valid Jinja2 template with 'input1' and 'input2' variables.
+    5. If 'resolution_prompt' is present, verifies it as a valid Jinja2 template with 'inputs' variable.
+    6. Optionally checks if 'model' is a string (if present).
+    7. Optionally checks 'blocking_keys' (if present, further checks are performed).
+
+    Raises:
+        ValueError: If required keys are missing, if templates are invalid or missing required variables,
+                    or if any other configuration aspect is incorrect or inconsistent.
+        TypeError: If the types of configuration values are incorrect, such as 'schema' not being a dict
+                   or 'model' not being a string.
+    """
+    required_keys = ["comparison_prompt", "output"]
+    for key in required_keys:
+        if key not in self.config:
+            raise ValueError(
+                f"Missing required key '{key}' in ResolveOperation configuration"
+            )
+
+    if "schema" not in self.config["output"]:
+        raise ValueError("Missing 'schema' in 'output' configuration")
+
+    if not isinstance(self.config["output"]["schema"], dict):
+        raise TypeError("'schema' in 'output' configuration must be a dictionary")
+
+    if not self.config["output"]["schema"]:
+        raise ValueError("'schema' in 'output' configuration cannot be empty")
+
+    # Check if the comparison_prompt is a valid Jinja2 template
+    try:
+        comparison_template = Template(self.config["comparison_prompt"])
+        comparison_vars = comparison_template.environment.parse(
+            self.config["comparison_prompt"]
+        ).find_all(jinja2.nodes.Name)
+        comparison_var_names = {var.name for var in comparison_vars}
+        if (
+            "input1" not in comparison_var_names
+            or "input2" not in comparison_var_names
+        ):
+            raise ValueError(
+                "'comparison_prompt' must contain both 'input1' and 'input2' variables"
+            )
+
+        if "resolution_prompt" in self.config:
+            reduction_template = Template(self.config["resolution_prompt"])
+            reduction_vars = reduction_template.environment.parse(
+                self.config["resolution_prompt"]
+            ).find_all(jinja2.nodes.Name)
+            reduction_var_names = {var.name for var in reduction_vars}
+            if "inputs" not in reduction_var_names:
+                raise ValueError(
+                    "'resolution_prompt' must contain 'inputs' variable"
+                )
+    except Exception as e:
+        raise ValueError(f"Invalid Jinja2 template: {str(e)}")
+
+    # Check if the model is specified (optional)
+    if "model" in self.config and not isinstance(self.config["model"], str):
+        raise TypeError("'model' in configuration must be a string")
+
+    # Check blocking_keys (optional)
+    if "blocking_keys" in self.config:
+        if not isinstance(self.config["blocking_keys"], list):
+            raise TypeError("'blocking_keys' must be a list")
+        if not all(isinstance(key, str) for key in self.config["blocking_keys"]):
+            raise TypeError("All items in 'blocking_keys' must be strings")
+
+    # Check blocking_threshold (optional)
+    if "blocking_threshold" in self.config:
+        if not isinstance(self.config["blocking_threshold"], (int, float)):
+            raise TypeError("'blocking_threshold' must be a number")
+        if not 0 <= self.config["blocking_threshold"] <= 1:
+            raise ValueError("'blocking_threshold' must be between 0 and 1")
+
+    # Check blocking_conditions (optional)
+    if "blocking_conditions" in self.config:
+        if not isinstance(self.config["blocking_conditions"], list):
+            raise TypeError("'blocking_conditions' must be a list")
+        if not all(
+            isinstance(cond, str) for cond in self.config["blocking_conditions"]
+        ):
+            raise TypeError("All items in 'blocking_conditions' must be strings")
+
+    # Check if input schema is provided and valid (optional)
+    if "input" in self.config:
+        if "schema" not in self.config["input"]:
+            raise ValueError("Missing 'schema' in 'input' configuration")
+        if not isinstance(self.config["input"]["schema"], dict):
+            raise TypeError(
+                "'schema' in 'input' configuration must be a dictionary"
+            )
+
+    # Check limit_comparisons (optional)
+    if "limit_comparisons" in self.config:
+        if not isinstance(self.config["limit_comparisons"], int):
+            raise TypeError("'limit_comparisons' must be an integer")
+        if self.config["limit_comparisons"] <= 0:
+            raise ValueError("'limit_comparisons' must be a positive integer")
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.operations.reduce.ReduceOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + +

A class that implements a reduce operation on input data using language models.

+

This class extends BaseOperation to provide functionality for reducing grouped data +using various strategies including batch reduce, incremental reduce, and parallel fold and merge.

+ + + + + + +
+ Source code in docetl/operations/reduce.py +
  31
+  32
+  33
+  34
+  35
+  36
+  37
+  38
+  39
+  40
+  41
+  42
+  43
+  44
+  45
+  46
+  47
+  48
+  49
+  50
+  51
+  52
+  53
+  54
+  55
+  56
+  57
+  58
+  59
+  60
+  61
+  62
+  63
+  64
+  65
+  66
+  67
+  68
+  69
+  70
+  71
+  72
+  73
+  74
+  75
+  76
+  77
+  78
+  79
+  80
+  81
+  82
+  83
+  84
+  85
+  86
+  87
+  88
+  89
+  90
+  91
+  92
+  93
+  94
+  95
+  96
+  97
+  98
+  99
+ 100
+ 101
+ 102
+ 103
+ 104
+ 105
+ 106
+ 107
+ 108
+ 109
+ 110
+ 111
+ 112
+ 113
+ 114
+ 115
+ 116
+ 117
+ 118
+ 119
+ 120
+ 121
+ 122
+ 123
+ 124
+ 125
+ 126
+ 127
+ 128
+ 129
+ 130
+ 131
+ 132
+ 133
+ 134
+ 135
+ 136
+ 137
+ 138
+ 139
+ 140
+ 141
+ 142
+ 143
+ 144
+ 145
+ 146
+ 147
+ 148
+ 149
+ 150
+ 151
+ 152
+ 153
+ 154
+ 155
+ 156
+ 157
+ 158
+ 159
+ 160
+ 161
+ 162
+ 163
+ 164
+ 165
+ 166
+ 167
+ 168
+ 169
+ 170
+ 171
+ 172
+ 173
+ 174
+ 175
+ 176
+ 177
+ 178
+ 179
+ 180
+ 181
+ 182
+ 183
+ 184
+ 185
+ 186
+ 187
+ 188
+ 189
+ 190
+ 191
+ 192
+ 193
+ 194
+ 195
+ 196
+ 197
+ 198
+ 199
+ 200
+ 201
+ 202
+ 203
+ 204
+ 205
+ 206
+ 207
+ 208
+ 209
+ 210
+ 211
+ 212
+ 213
+ 214
+ 215
+ 216
+ 217
+ 218
+ 219
+ 220
+ 221
+ 222
+ 223
+ 224
+ 225
+ 226
+ 227
+ 228
+ 229
+ 230
+ 231
+ 232
+ 233
+ 234
+ 235
+ 236
+ 237
+ 238
+ 239
+ 240
+ 241
+ 242
+ 243
+ 244
+ 245
+ 246
+ 247
+ 248
+ 249
+ 250
+ 251
+ 252
+ 253
+ 254
+ 255
+ 256
+ 257
+ 258
+ 259
+ 260
+ 261
+ 262
+ 263
+ 264
+ 265
+ 266
+ 267
+ 268
+ 269
+ 270
+ 271
+ 272
+ 273
+ 274
+ 275
+ 276
+ 277
+ 278
+ 279
+ 280
+ 281
+ 282
+ 283
+ 284
+ 285
+ 286
+ 287
+ 288
+ 289
+ 290
+ 291
+ 292
+ 293
+ 294
+ 295
+ 296
+ 297
+ 298
+ 299
+ 300
+ 301
+ 302
+ 303
+ 304
+ 305
+ 306
+ 307
+ 308
+ 309
+ 310
+ 311
+ 312
+ 313
+ 314
+ 315
+ 316
+ 317
+ 318
+ 319
+ 320
+ 321
+ 322
+ 323
+ 324
+ 325
+ 326
+ 327
+ 328
+ 329
+ 330
+ 331
+ 332
+ 333
+ 334
+ 335
+ 336
+ 337
+ 338
+ 339
+ 340
+ 341
+ 342
+ 343
+ 344
+ 345
+ 346
+ 347
+ 348
+ 349
+ 350
+ 351
+ 352
+ 353
+ 354
+ 355
+ 356
+ 357
+ 358
+ 359
+ 360
+ 361
+ 362
+ 363
+ 364
+ 365
+ 366
+ 367
+ 368
+ 369
+ 370
+ 371
+ 372
+ 373
+ 374
+ 375
+ 376
+ 377
+ 378
+ 379
+ 380
+ 381
+ 382
+ 383
+ 384
+ 385
+ 386
+ 387
+ 388
+ 389
+ 390
+ 391
+ 392
+ 393
+ 394
+ 395
+ 396
+ 397
+ 398
+ 399
+ 400
+ 401
+ 402
+ 403
+ 404
+ 405
+ 406
+ 407
+ 408
+ 409
+ 410
+ 411
+ 412
+ 413
+ 414
+ 415
+ 416
+ 417
+ 418
+ 419
+ 420
+ 421
+ 422
+ 423
+ 424
+ 425
+ 426
+ 427
+ 428
+ 429
+ 430
+ 431
+ 432
+ 433
+ 434
+ 435
+ 436
+ 437
+ 438
+ 439
+ 440
+ 441
+ 442
+ 443
+ 444
+ 445
+ 446
+ 447
+ 448
+ 449
+ 450
+ 451
+ 452
+ 453
+ 454
+ 455
+ 456
+ 457
+ 458
+ 459
+ 460
+ 461
+ 462
+ 463
+ 464
+ 465
+ 466
+ 467
+ 468
+ 469
+ 470
+ 471
+ 472
+ 473
+ 474
+ 475
+ 476
+ 477
+ 478
+ 479
+ 480
+ 481
+ 482
+ 483
+ 484
+ 485
+ 486
+ 487
+ 488
+ 489
+ 490
+ 491
+ 492
+ 493
+ 494
+ 495
+ 496
+ 497
+ 498
+ 499
+ 500
+ 501
+ 502
+ 503
+ 504
+ 505
+ 506
+ 507
+ 508
+ 509
+ 510
+ 511
+ 512
+ 513
+ 514
+ 515
+ 516
+ 517
+ 518
+ 519
+ 520
+ 521
+ 522
+ 523
+ 524
+ 525
+ 526
+ 527
+ 528
+ 529
+ 530
+ 531
+ 532
+ 533
+ 534
+ 535
+ 536
+ 537
+ 538
+ 539
+ 540
+ 541
+ 542
+ 543
+ 544
+ 545
+ 546
+ 547
+ 548
+ 549
+ 550
+ 551
+ 552
+ 553
+ 554
+ 555
+ 556
+ 557
+ 558
+ 559
+ 560
+ 561
+ 562
+ 563
+ 564
+ 565
+ 566
+ 567
+ 568
+ 569
+ 570
+ 571
+ 572
+ 573
+ 574
+ 575
+ 576
+ 577
+ 578
+ 579
+ 580
+ 581
+ 582
+ 583
+ 584
+ 585
+ 586
+ 587
+ 588
+ 589
+ 590
+ 591
+ 592
+ 593
+ 594
+ 595
+ 596
+ 597
+ 598
+ 599
+ 600
+ 601
+ 602
+ 603
+ 604
+ 605
+ 606
+ 607
+ 608
+ 609
+ 610
+ 611
+ 612
+ 613
+ 614
+ 615
+ 616
+ 617
+ 618
+ 619
+ 620
+ 621
+ 622
+ 623
+ 624
+ 625
+ 626
+ 627
+ 628
+ 629
+ 630
+ 631
+ 632
+ 633
+ 634
+ 635
+ 636
+ 637
+ 638
+ 639
+ 640
+ 641
+ 642
+ 643
+ 644
+ 645
+ 646
+ 647
+ 648
+ 649
+ 650
+ 651
+ 652
+ 653
+ 654
+ 655
+ 656
+ 657
+ 658
+ 659
+ 660
+ 661
+ 662
+ 663
+ 664
+ 665
+ 666
+ 667
+ 668
+ 669
+ 670
+ 671
+ 672
+ 673
+ 674
+ 675
+ 676
+ 677
+ 678
+ 679
+ 680
+ 681
+ 682
+ 683
+ 684
+ 685
+ 686
+ 687
+ 688
+ 689
+ 690
+ 691
+ 692
+ 693
+ 694
+ 695
+ 696
+ 697
+ 698
+ 699
+ 700
+ 701
+ 702
+ 703
+ 704
+ 705
+ 706
+ 707
+ 708
+ 709
+ 710
+ 711
+ 712
+ 713
+ 714
+ 715
+ 716
+ 717
+ 718
+ 719
+ 720
+ 721
+ 722
+ 723
+ 724
+ 725
+ 726
+ 727
+ 728
+ 729
+ 730
+ 731
+ 732
+ 733
+ 734
+ 735
+ 736
+ 737
+ 738
+ 739
+ 740
+ 741
+ 742
+ 743
+ 744
+ 745
+ 746
+ 747
+ 748
+ 749
+ 750
+ 751
+ 752
+ 753
+ 754
+ 755
+ 756
+ 757
+ 758
+ 759
+ 760
+ 761
+ 762
+ 763
+ 764
+ 765
+ 766
+ 767
+ 768
+ 769
+ 770
+ 771
+ 772
+ 773
+ 774
+ 775
+ 776
+ 777
+ 778
+ 779
+ 780
+ 781
+ 782
+ 783
+ 784
+ 785
+ 786
+ 787
+ 788
+ 789
+ 790
+ 791
+ 792
+ 793
+ 794
+ 795
+ 796
+ 797
+ 798
+ 799
+ 800
+ 801
+ 802
+ 803
+ 804
+ 805
+ 806
+ 807
+ 808
+ 809
+ 810
+ 811
+ 812
+ 813
+ 814
+ 815
+ 816
+ 817
+ 818
+ 819
+ 820
+ 821
+ 822
+ 823
+ 824
+ 825
+ 826
+ 827
+ 828
+ 829
+ 830
+ 831
+ 832
+ 833
+ 834
+ 835
+ 836
+ 837
+ 838
+ 839
+ 840
+ 841
+ 842
+ 843
+ 844
+ 845
+ 846
+ 847
+ 848
+ 849
+ 850
+ 851
+ 852
+ 853
+ 854
+ 855
+ 856
+ 857
+ 858
+ 859
+ 860
+ 861
+ 862
+ 863
+ 864
+ 865
+ 866
+ 867
+ 868
+ 869
+ 870
+ 871
+ 872
+ 873
+ 874
+ 875
+ 876
+ 877
+ 878
+ 879
+ 880
+ 881
+ 882
+ 883
+ 884
+ 885
+ 886
+ 887
+ 888
+ 889
+ 890
+ 891
+ 892
+ 893
+ 894
+ 895
+ 896
+ 897
+ 898
+ 899
+ 900
+ 901
+ 902
+ 903
+ 904
+ 905
+ 906
+ 907
+ 908
+ 909
+ 910
+ 911
+ 912
+ 913
+ 914
+ 915
+ 916
+ 917
+ 918
+ 919
+ 920
+ 921
+ 922
+ 923
+ 924
+ 925
+ 926
+ 927
+ 928
+ 929
+ 930
+ 931
+ 932
+ 933
+ 934
+ 935
+ 936
+ 937
+ 938
+ 939
+ 940
+ 941
+ 942
+ 943
+ 944
+ 945
+ 946
+ 947
+ 948
+ 949
+ 950
+ 951
+ 952
+ 953
+ 954
+ 955
+ 956
+ 957
+ 958
+ 959
+ 960
+ 961
+ 962
+ 963
+ 964
+ 965
+ 966
+ 967
+ 968
+ 969
+ 970
+ 971
+ 972
+ 973
+ 974
+ 975
+ 976
+ 977
+ 978
+ 979
+ 980
+ 981
+ 982
+ 983
+ 984
+ 985
+ 986
+ 987
+ 988
+ 989
+ 990
+ 991
+ 992
+ 993
+ 994
+ 995
+ 996
+ 997
+ 998
+ 999
+1000
+1001
+1002
+1003
+1004
+1005
+1006
class ReduceOperation(BaseOperation):
+    """
+    A class that implements a reduce operation on input data using language models.
+
+    This class extends BaseOperation to provide functionality for reducing grouped data
+    using various strategies including batch reduce, incremental reduce, and parallel fold and merge.
+    """
+
+    class schema(BaseOperation.schema):
+        type: str = "reduce"
+        reduce_key: Union[str, List[str]]
+        output: Optional[Dict[str, Any]] = None
+        prompt: Optional[str] = None
+        optimize: Optional[bool] = None
+        synthesize_resolve: Optional[bool] = None
+        model: Optional[str] = None
+        input: Optional[Dict[str, Any]] = None
+        pass_through: Optional[bool] = None
+        associative: Optional[bool] = None
+        fold_prompt: Optional[str] = None
+        fold_batch_size: Optional[int] = None
+        value_sampling: Optional[Dict[str, Any]] = None
+        verbose: Optional[bool] = None
+        timeout: Optional[int] = None
+        litellm_completion_kwargs: Dict[str, Any] = Field(default_factory=dict)
+        enable_observability: bool = False
+
+    def __init__(self, *args, **kwargs):
+        """
+        Initialize the ReduceOperation.
+
+        Args:
+            *args: Variable length argument list.
+            **kwargs: Arbitrary keyword arguments.
+        """
+        super().__init__(*args, **kwargs)
+        self.min_samples = 5
+        self.max_samples = 1000
+        self.fold_times = deque(maxlen=self.max_samples)
+        self.merge_times = deque(maxlen=self.max_samples)
+        self.lock = Lock()
+        self.config["reduce_key"] = (
+            [self.config["reduce_key"]]
+            if isinstance(self.config["reduce_key"], str)
+            else self.config["reduce_key"]
+        )
+        self.intermediates = {}
+        self.lineage_keys = self.config.get("output", {}).get("lineage", [])
+
+    def syntax_check(self) -> None:
+        """
+        Perform comprehensive syntax checks on the configuration of the ReduceOperation.
+
+        This method validates the presence and correctness of all required configuration keys, Jinja2 templates, and ensures the correct
+        structure and types of the entire configuration.
+
+        The method performs the following checks:
+        1. Verifies the presence of all required keys in the configuration.
+        2. Validates the structure and content of the 'output' configuration, including its 'schema'.
+        3. Checks if the main 'prompt' is a valid Jinja2 template and contains the required 'inputs' variable.
+        4. If 'merge_prompt' is specified, ensures that 'fold_prompt' is also present.
+        5. If 'fold_prompt' is present, verifies the existence of 'fold_batch_size'.
+        6. Validates the 'fold_prompt' as a Jinja2 template with required variables 'inputs' and 'output'.
+        7. If present, checks 'merge_prompt' as a valid Jinja2 template with required 'outputs' variable.
+        8. Verifies types of various configuration inputs (e.g., 'fold_batch_size' as int).
+        9. Checks for the presence and validity of optional configurations like 'model'.
+
+        Raises:
+            ValueError: If any required configuration is missing, if templates are invalid or missing required
+                        variables, or if any other configuration aspect is incorrect or inconsistent.
+            TypeError: If any configuration value has an incorrect type, such as 'schema' not being a dict
+                       or 'fold_batch_size' not being an integer.
+        """
+        required_keys = ["reduce_key", "prompt", "output"]
+        for key in required_keys:
+            if key not in self.config:
+                raise ValueError(
+                    f"Missing required key '{key}' in {self.config['name']} configuration"
+                )
+
+        if "schema" not in self.config["output"]:
+            raise ValueError(
+                f"Missing 'schema' in {self.config['name']} 'output' configuration"
+            )
+
+        if not isinstance(self.config["output"]["schema"], dict):
+            raise TypeError(
+                f"'schema' in {self.config['name']} 'output' configuration must be a dictionary"
+            )
+
+        if not self.config["output"]["schema"]:
+            raise ValueError(
+                f"'schema' in {self.config['name']} 'output' configuration cannot be empty"
+            )
+
+        # Check if the prompt is a valid Jinja2 template
+        try:
+            template = Template(self.config["prompt"])
+            template_vars = template.environment.parse(self.config["prompt"]).find_all(
+                jinja2.nodes.Name
+            )
+            template_var_names = {var.name for var in template_vars}
+            if "inputs" not in template_var_names:
+                raise ValueError(
+                    f"Prompt template for {self.config['name']} must include the 'inputs' variable"
+                )
+        except Exception as e:
+            raise ValueError(
+                f"Invalid Jinja2 template in {self.config['name']} 'prompt': {str(e)}"
+            )
+
+        # Check if fold_prompt is a valid Jinja2 template (now required if merge exists)
+        if "merge_prompt" in self.config:
+            if "fold_prompt" not in self.config:
+                raise ValueError(
+                    f"'fold_prompt' is required when 'merge_prompt' is specified in {self.config['name']}"
+                )
+
+        if "fold_prompt" in self.config:
+            if "fold_batch_size" not in self.config:
+                raise ValueError(
+                    f"'fold_batch_size' is required when 'fold_prompt' is specified in {self.config['name']}"
+                )
+
+            try:
+                fold_template = Template(self.config["fold_prompt"])
+                fold_template_vars = fold_template.environment.parse(
+                    self.config["fold_prompt"]
+                ).find_all(jinja2.nodes.Name)
+                fold_template_var_names = {var.name for var in fold_template_vars}
+                required_vars = {"inputs", "output"}
+                if not required_vars.issubset(fold_template_var_names):
+                    raise ValueError(
+                        f"Fold template in {self.config['name']} must include variables: {required_vars}. Current template includes: {fold_template_var_names}"
+                    )
+            except Exception as e:
+                raise ValueError(
+                    f"Invalid Jinja2 template in {self.config['name']} 'fold_prompt': {str(e)}"
+                )
+
+        # Check merge_prompt and merge_batch_size
+        if "merge_prompt" in self.config:
+            if "merge_batch_size" not in self.config:
+                raise ValueError(
+                    f"'merge_batch_size' is required when 'merge_prompt' is specified in {self.config['name']}"
+                )
+
+            try:
+                merge_template = Template(self.config["merge_prompt"])
+                merge_template_vars = merge_template.environment.parse(
+                    self.config["merge_prompt"]
+                ).find_all(jinja2.nodes.Name)
+                merge_template_var_names = {var.name for var in merge_template_vars}
+                if "outputs" not in merge_template_var_names:
+                    raise ValueError(
+                        f"Merge template in {self.config['name']} must include the 'outputs' variable"
+                    )
+            except Exception as e:
+                raise ValueError(
+                    f"Invalid Jinja2 template in {self.config['name']} 'merge_prompt': {str(e)}"
+                )
+
+        # Check if the model is specified (optional)
+        if "model" in self.config and not isinstance(self.config["model"], str):
+            raise TypeError(
+                f"'model' in {self.config['name']} configuration must be a string"
+            )
+
+        # Check if reduce_key is a string or a list of strings
+        if not isinstance(self.config["reduce_key"], (str, list)):
+            raise TypeError(
+                f"'reduce_key' in {self.config['name']} configuration must be a string or a list of strings"
+            )
+        if isinstance(self.config["reduce_key"], list):
+            if not all(isinstance(key, str) for key in self.config["reduce_key"]):
+                raise TypeError(
+                    f"All elements in 'reduce_key' list in {self.config['name']} configuration must be strings"
+                )
+
+        # Check if input schema is provided and valid (optional)
+        if "input" in self.config:
+            if "schema" not in self.config["input"]:
+                raise ValueError(
+                    f"Missing 'schema' in {self.config['name']} 'input' configuration"
+                )
+            if not isinstance(self.config["input"]["schema"], dict):
+                raise TypeError(
+                    f"'schema' in {self.config['name']} 'input' configuration must be a dictionary"
+                )
+
+        # Check if fold_batch_size and merge_batch_size are positive integers
+        for key in ["fold_batch_size", "merge_batch_size"]:
+            if key in self.config:
+                if not isinstance(self.config[key], int) or self.config[key] <= 0:
+                    raise ValueError(
+                        f"'{key}' in {self.config['name']} configuration must be a positive integer"
+                    )
+
+        if "value_sampling" in self.config:
+            sampling = self.config["value_sampling"]
+            if not isinstance(sampling, dict):
+                raise TypeError(
+                    f"'value_sampling' in {self.config['name']} configuration must be a dictionary"
+                )
+
+            if "enabled" not in sampling:
+                raise ValueError(
+                    f"'enabled' is required in {self.config['name']} 'value_sampling' configuration"
+                )
+            if not isinstance(sampling["enabled"], bool):
+                raise TypeError(
+                    f"'enabled' in {self.config['name']} 'value_sampling' configuration must be a boolean"
+                )
+
+            if sampling["enabled"]:
+                if "sample_size" not in sampling:
+                    raise ValueError(
+                        f"'sample_size' is required when value_sampling is enabled in {self.config['name']}"
+                    )
+                if (
+                    not isinstance(sampling["sample_size"], int)
+                    or sampling["sample_size"] <= 0
+                ):
+                    raise ValueError(
+                        f"'sample_size' in {self.config['name']} configuration must be a positive integer"
+                    )
+
+                if "method" not in sampling:
+                    raise ValueError(
+                        f"'method' is required when value_sampling is enabled in {self.config['name']}"
+                    )
+                if sampling["method"] not in [
+                    "random",
+                    "first_n",
+                    "cluster",
+                    "sem_sim",
+                ]:
+                    raise ValueError(
+                        f"Invalid 'method'. Must be 'random', 'first_n', or 'embedding' in {self.config['name']}"
+                    )
+
+                if sampling["method"] == "embedding":
+                    if "embedding_model" not in sampling:
+                        raise ValueError(
+                            f"'embedding_model' is required when using embedding-based sampling in {self.config['name']}"
+                        )
+                    if "embedding_keys" not in sampling:
+                        raise ValueError(
+                            f"'embedding_keys' is required when using embedding-based sampling in {self.config['name']}"
+                        )
+
+        # Check if lineage is a list of strings
+        if "lineage" in self.config.get("output", {}):
+            if not isinstance(self.config["output"]["lineage"], list):
+                raise TypeError(
+                    f"'lineage' in {self.config['name']} 'output' configuration must be a list"
+                )
+            if not all(
+                isinstance(key, str) for key in self.config["output"]["lineage"]
+            ):
+                raise TypeError(
+                    f"All elements in 'lineage' list in {self.config['name']} 'output' configuration must be strings"
+                )
+
+        self.gleaning_check()
+
+    def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+        """
+        Execute the reduce operation on the provided input data.
+
+        This method sorts and groups the input data by the reduce key(s), then processes each group
+        using either parallel fold and merge, incremental reduce, or batch reduce strategies.
+
+        Args:
+            input_data (List[Dict]): The input data to process.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.
+        """
+        if self.config.get("gleaning", {}).get("validation_prompt", None):
+            self.console.log(
+                f"Using gleaning with validation prompt: {self.config.get('gleaning', {}).get('validation_prompt', '')}"
+            )
+
+        reduce_keys = self.config["reduce_key"]
+        if isinstance(reduce_keys, str):
+            reduce_keys = [reduce_keys]
+        input_schema = self.config.get("input", {}).get("schema", {})
+
+        if self.status:
+            self.status.stop()
+
+        # Check if we need to group everything into one group
+        if reduce_keys == ["_all"] or reduce_keys == "_all":
+            grouped_data = [("_all", input_data)]
+        else:
+            # Group the input data by the reduce key(s) while maintaining original order
+            def get_group_key(item):
+                key_values = []
+                for key in reduce_keys:
+                    value = item[key]
+                    # Special handling for list-type values
+                    if isinstance(value, list):
+                        key_values.append(tuple(sorted(value)))  # Convert list to sorted tuple
+                    else:
+                        key_values.append(value)
+                return tuple(key_values)
+
+            grouped_data = {}
+            for item in input_data:
+                key = get_group_key(item)
+                if key not in grouped_data:
+                    grouped_data[key] = []
+                grouped_data[key].append(item)
+
+            # Convert the grouped data to a list of tuples
+            grouped_data = list(grouped_data.items())
+
+        def process_group(
+            key: Tuple, group_elems: List[Dict]
+        ) -> Tuple[Optional[Dict], float]:
+            if input_schema:
+                group_list = [
+                    {k: item[k] for k in input_schema.keys() if k in item}
+                    for item in group_elems
+                ]
+            else:
+                group_list = group_elems
+
+            total_cost = 0.0
+
+            # Apply value sampling if enabled
+            value_sampling = self.config.get("value_sampling", {})
+            if value_sampling.get("enabled", False):
+                sample_size = min(value_sampling["sample_size"], len(group_list))
+                method = value_sampling["method"]
+
+                if method == "random":
+                    group_sample = random.sample(group_list, sample_size)
+                    group_sample.sort(key=lambda x: group_list.index(x))
+                elif method == "first_n":
+                    group_sample = group_list[:sample_size]
+                elif method == "cluster":
+                    group_sample, embedding_cost = self._cluster_based_sampling(
+                        group_list, value_sampling, sample_size
+                    )
+                    group_sample.sort(key=lambda x: group_list.index(x))
+                    total_cost += embedding_cost
+                elif method == "sem_sim":
+                    group_sample, embedding_cost = self._semantic_similarity_sampling(
+                        key, group_list, value_sampling, sample_size
+                    )
+                    group_sample.sort(key=lambda x: group_list.index(x))
+                    total_cost += embedding_cost
+
+                group_list = group_sample
+
+            # Only execute merge-based plans if associative = True
+            if "merge_prompt" in self.config and self.config.get("associative", True):
+                result, prompts, cost = self._parallel_fold_and_merge(key, group_list)
+            elif (
+                self.config.get("fold_batch_size", None)
+                and self.config.get("fold_batch_size") >= len(group_list)
+            ):
+                # If the fold batch size is greater than or equal to the number of items in the group,
+                # we can just run a single fold operation
+                result, prompt, cost = self._batch_reduce(key, group_list)
+                prompts = [prompt]
+            elif "fold_prompt" in self.config:
+                result, prompts, cost = self._incremental_reduce(key, group_list)
+            else:
+                result, prompt, cost = self._batch_reduce(key, group_list)
+                prompts = [prompt]
+
+            total_cost += cost
+
+            # Add the counts of items in the group to the result
+            result[f"_counts_prereduce_{self.config['name']}"] = len(group_elems)
+
+            if self.config.get("enable_observability", False):
+                # Add the _observability_{self.config['name']} key to the result
+                result[f"_observability_{self.config['name']}"] = {
+                    "prompts": prompts
+                }
+
+            # Apply pass-through at the group level
+            if (
+                result is not None
+                and self.config.get("pass_through", False)
+                and group_elems
+            ):
+                for k, v in group_elems[0].items():
+                    if k not in self.config["output"]["schema"] and k not in result:
+                        result[k] = v
+
+            # Add lineage information
+            if result is not None and self.lineage_keys:
+                lineage = []
+                for item in group_elems:
+                    lineage_item = {
+                        k: item.get(k) for k in self.lineage_keys if k in item
+                    }
+                    if lineage_item:
+                        lineage.append(lineage_item)
+                result[f"{self.config['name']}_lineage"] = lineage
+
+            return result, total_cost
+
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            futures = [
+                executor.submit(process_group, key, group)
+                for key, group in grouped_data
+            ]
+            results = []
+            total_cost = 0
+            for future in rich_as_completed(
+                futures,
+                total=len(futures),
+                desc=f"Processing {self.config['name']} (reduce) on all documents",
+                leave=True,
+                console=self.console,
+            ):
+                output, item_cost = future.result()
+                total_cost += item_cost
+                if output is not None:
+                    results.append(output)
+
+        if self.config.get("persist_intermediates", False):
+            for result in results:
+                key = tuple(result[k] for k in self.config["reduce_key"])
+                if key in self.intermediates:
+                    result[f"_{self.config['name']}_intermediates"] = (
+                        self.intermediates[key]
+                    )
+
+        if self.status:
+            self.status.start()
+
+        return results, total_cost
+
+    def _cluster_based_sampling(
+        self, group_list: List[Dict], value_sampling: Dict, sample_size: int
+    ) -> Tuple[List[Dict], float]:
+        if sample_size >= len(group_list):
+            return group_list, 0
+
+        clusters, cost = cluster_documents(
+            group_list, value_sampling, sample_size, self.runner.api
+        )
+
+        sampled_items = []
+        idx_added_already = set()
+        num_clusters = len(clusters)
+        for i in range(sample_size):
+            # Add a random item from the cluster
+            idx = i % num_clusters
+
+            # Skip if there are no items in the cluster
+            if len(clusters[idx]) == 0:
+                continue
+
+            if len(clusters[idx]) == 1:
+                # If there's only one item in the cluster, add it directly if we haven't already
+                if idx not in idx_added_already:
+                    sampled_items.append(clusters[idx][0])
+                continue
+
+            random_choice_idx = random.randint(0, len(clusters[idx]) - 1)
+            max_attempts = 10
+            while random_choice_idx in idx_added_already and max_attempts > 0:
+                random_choice_idx = random.randint(0, len(clusters[idx]) - 1)
+                max_attempts -= 1
+            idx_added_already.add(random_choice_idx)
+            sampled_items.append(clusters[idx][random_choice_idx])
+
+        return sampled_items, cost
+
+    def _semantic_similarity_sampling(
+        self, key: Tuple, group_list: List[Dict], value_sampling: Dict, sample_size: int
+    ) -> Tuple[List[Dict], float]:
+        embedding_model = value_sampling["embedding_model"]
+        query_text_template = Template(value_sampling["query_text"])
+        query_text = query_text_template.render(
+            reduce_key=dict(zip(self.config["reduce_key"], key))
+        )
+
+        embeddings, cost = get_embeddings_for_clustering(
+            group_list, value_sampling, self.runner.api
+        )
+
+        query_response = self.runner.api.gen_embedding(embedding_model, [query_text])
+        query_embedding = query_response["data"][0]["embedding"]
+        cost += completion_cost(query_response)
+
+        from sklearn.metrics.pairwise import cosine_similarity
+
+        similarities = cosine_similarity([query_embedding], embeddings)[0]
+
+        top_k_indices = np.argsort(similarities)[-sample_size:]
+
+        return [group_list[i] for i in top_k_indices], cost
+
+    def _parallel_fold_and_merge(
+        self, key: Tuple, group_list: List[Dict]
+    ) -> Tuple[Optional[Dict], float]:
+        """
+        Perform parallel folding and merging on a group of items.
+
+        This method implements a strategy that combines parallel folding of input items
+        and merging of intermediate results to efficiently process large groups. It works as follows:
+        1. The input group is initially divided into smaller batches for efficient processing.
+        2. The method performs an initial round of folding operations on these batches.
+        3. After the first round of folds, a few merges are performed to estimate the merge runtime.
+        4. Based on the estimated merge runtime and observed fold runtime, it calculates the optimal number of parallel folds. Subsequent rounds of folding are then performed concurrently, with the number of parallel folds determined by the runtime estimates.
+        5. The folding process repeats in rounds, progressively reducing the number of items to be processed.
+        6. Once all folding operations are complete, the method recursively performs final merges on the fold results to combine them into a final result.
+        7. Throughout this process, the method may adjust the number of parallel folds based on updated performance metrics (i.e., fold and merge runtimes) to maintain efficiency.
+
+        Args:
+            key (Tuple): The reduce key tuple for the group.
+            group_list (List[Dict]): The list of items in the group to be processed.
+
+        Returns:
+            Tuple[Optional[Dict], float]: A tuple containing the final merged result (or None if processing failed)
+            and the total cost of the operation.
+        """
+        fold_batch_size = self.config["fold_batch_size"]
+        merge_batch_size = self.config["merge_batch_size"]
+        total_cost = 0
+        prompts = []
+        def calculate_num_parallel_folds():
+            fold_time, fold_default = self.get_fold_time()
+            merge_time, merge_default = self.get_merge_time()
+            num_group_items = len(group_list)
+            return (
+                max(
+                    1,
+                    int(
+                        (fold_time * num_group_items * math.log(merge_batch_size))
+                        / (fold_batch_size * merge_time)
+                    ),
+                ),
+                fold_default or merge_default,
+            )
+
+        num_parallel_folds, used_default_times = calculate_num_parallel_folds()
+        fold_results = []
+        remaining_items = group_list
+
+        if self.config.get("persist_intermediates", False):
+            self.intermediates[key] = []
+            iter_count = 0
+
+        # Parallel folding and merging
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            while remaining_items:
+                # Folding phase
+                fold_futures = []
+                for i in range(min(num_parallel_folds, len(remaining_items))):
+                    batch = remaining_items[:fold_batch_size]
+                    remaining_items = remaining_items[fold_batch_size:]
+                    current_output = fold_results[i] if i < len(fold_results) else None
+                    fold_futures.append(
+                        executor.submit(
+                            self._increment_fold, key, batch, current_output
+                        )
+                    )
+
+                new_fold_results = []
+                for future in as_completed(fold_futures):
+                    result, prompt, cost = future.result()
+                    total_cost += cost
+                    prompts.append(prompt)
+                    if result is not None:
+                        new_fold_results.append(result)
+                        if self.config.get("persist_intermediates", False):
+                            self.intermediates[key].append(
+                                {
+                                    "iter": iter_count,
+                                    "intermediate": result,
+                                    "scratchpad": result["updated_scratchpad"],
+                                }
+                            )
+                            iter_count += 1
+
+                # Update fold_results with new results
+                fold_results = new_fold_results + fold_results[len(new_fold_results) :]
+
+                # Single pass merging phase
+                if (
+                    len(self.merge_times) < self.min_samples
+                    and len(fold_results) >= merge_batch_size
+                ):
+                    merge_futures = []
+                    for i in range(0, len(fold_results), merge_batch_size):
+                        batch = fold_results[i : i + merge_batch_size]
+                        merge_futures.append(
+                            executor.submit(self._merge_results, key, batch)
+                        )
+
+                    new_results = []
+                    for future in as_completed(merge_futures):
+                        result, prompt, cost = future.result()
+                        total_cost += cost
+                        prompts.append(prompt)
+                        if result is not None:
+                            new_results.append(result)
+                            if self.config.get("persist_intermediates", False):
+                                self.intermediates[key].append(
+                                    {
+                                        "iter": iter_count,
+                                        "intermediate": result,
+                                        "scratchpad": None,
+                                    }
+                                )
+                                iter_count += 1
+
+                    fold_results = new_results
+
+                # Recalculate num_parallel_folds if we used default times
+                if used_default_times:
+                    new_num_parallel_folds, used_default_times = (
+                        calculate_num_parallel_folds()
+                    )
+                    if not used_default_times:
+                        self.console.log(
+                            f"Recalculated num_parallel_folds from {num_parallel_folds} to {new_num_parallel_folds}"
+                        )
+                        num_parallel_folds = new_num_parallel_folds
+
+        # Final merging if needed
+        while len(fold_results) > 1:
+            self.console.log(f"Finished folding! Merging {len(fold_results)} items.")
+            with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+                merge_futures = []
+                for i in range(0, len(fold_results), merge_batch_size):
+                    batch = fold_results[i : i + merge_batch_size]
+                    merge_futures.append(
+                        executor.submit(self._merge_results, key, batch)
+                    )
+
+                new_results = []
+                for future in as_completed(merge_futures):
+                    result, prompt, cost = future.result()
+                    total_cost += cost
+                    prompts.append(prompt)
+                    if result is not None:
+                        new_results.append(result)
+                        if self.config.get("persist_intermediates", False):
+                            self.intermediates[key].append(
+                                {
+                                    "iter": iter_count,
+                                    "intermediate": result,
+                                    "scratchpad": None,
+                                }
+                            )
+                            iter_count += 1
+
+                fold_results = new_results
+
+        return (fold_results[0], prompts, total_cost) if fold_results else (None, prompts, total_cost)
+
+    def _incremental_reduce(
+        self, key: Tuple, group_list: List[Dict]
+    ) -> Tuple[Optional[Dict], List[str], float]:
+        """
+        Perform an incremental reduce operation on a group of items.
+
+        This method processes the group in batches, incrementally folding the results.
+
+        Args:
+            key (Tuple): The reduce key tuple for the group.
+            group_list (List[Dict]): The list of items in the group to be processed.
+
+        Returns:
+            Tuple[Optional[Dict], List[str], float]: A tuple containing the final reduced result (or None if processing failed),
+            the list of prompts used, and the total cost of the operation.
+        """
+        fold_batch_size = self.config["fold_batch_size"]
+        total_cost = 0
+        current_output = None
+        prompts = []
+
+        # Calculate and log the number of folds to be performed
+        num_folds = (len(group_list) + fold_batch_size - 1) // fold_batch_size
+
+        scratchpad = ""
+        if self.config.get("persist_intermediates", False):
+            self.intermediates[key] = []
+            iter_count = 0
+
+        for i in range(0, len(group_list), fold_batch_size):
+            # Log the current iteration and total number of folds
+            current_fold = i // fold_batch_size + 1
+            if self.config.get("verbose", False):
+                self.console.log(
+                    f"Processing fold {current_fold} of {num_folds} for group with key {key}"
+                )
+            batch = group_list[i : i + fold_batch_size]
+
+            folded_output, prompt, fold_cost = self._increment_fold(
+                key, batch, current_output, scratchpad
+            )
+            total_cost += fold_cost
+            prompts.append(prompt)
+
+            if folded_output is None:
+                continue
+
+            if self.config.get("persist_intermediates", False):
+                self.intermediates[key].append(
+                    {
+                        "iter": iter_count,
+                        "intermediate": folded_output,
+                        "scratchpad": folded_output["updated_scratchpad"],
+                    }
+                )
+                iter_count += 1
+
+            # Pop off updated_scratchpad
+            if "updated_scratchpad" in folded_output:
+                scratchpad = folded_output["updated_scratchpad"]
+                if self.config.get("verbose", False):
+                    self.console.log(
+                        f"Updated scratchpad for fold {current_fold}: {scratchpad}"
+                    )
+                del folded_output["updated_scratchpad"]
+
+            current_output = folded_output
+
+        return current_output, prompts, total_cost
+
+    def validation_fn(self, response: Dict[str, Any]):
+        output = self.runner.api.parse_llm_response(
+            response,
+            schema=self.config["output"]["schema"],
+        )[0]
+        if self.runner.api.validate_output(self.config, output, self.console):
+            return output, True
+        return output, False
+
+    def _increment_fold(
+        self,
+        key: Tuple,
+        batch: List[Dict],
+        current_output: Optional[Dict],
+        scratchpad: Optional[str] = None,
+    ) -> Tuple[Optional[Dict], str, float]:
+        """
+        Perform an incremental fold operation on a batch of items.
+
+        This method folds a batch of items into the current output using the fold prompt.
+
+        Args:
+            key (Tuple): The reduce key tuple for the group.
+            batch (List[Dict]): The batch of items to be folded.
+            current_output (Optional[Dict]): The current accumulated output, if any.
+            scratchpad (Optional[str]): The scratchpad to use for the fold operation.
+        Returns:
+            Tuple[Optional[Dict], str, float]: A tuple containing the folded output (or None if processing failed),
+            the prompt used, and the cost of the fold operation.
+        """
+        if current_output is None:
+            return self._batch_reduce(key, batch, scratchpad)
+
+        start_time = time.time()
+        fold_prompt_template = Template(self.config["fold_prompt"])
+        fold_prompt = fold_prompt_template.render(
+            inputs=batch,
+            output=current_output,
+            reduce_key=dict(zip(self.config["reduce_key"], key)),
+        )
+
+        response = self.runner.api.call_llm(
+            self.config.get("model", self.default_model),
+            "reduce",
+            [{"role": "user", "content": fold_prompt}],
+            self.config["output"]["schema"],
+            scratchpad=scratchpad,
+            timeout_seconds=self.config.get("timeout", 120),
+            max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+            validation_config=(
+                {
+                    "num_retries": self.num_retries_on_validate_failure,
+                    "val_rule": self.config.get("validate", []),
+                    "validation_fn": self.validation_fn,
+                }
+                if self.config.get("validate", None)
+                else None
+            ),
+            bypass_cache=self.config.get("bypass_cache", False),
+            verbose=self.config.get("verbose", False),
+            litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+        )
+
+        end_time = time.time()
+        self._update_fold_time(end_time - start_time)
+
+        if response.validated:
+            folded_output = self.runner.api.parse_llm_response(
+                response.response,
+                schema=self.config["output"]["schema"],
+                manually_fix_errors=self.manually_fix_errors,
+            )[0]
+
+            folded_output.update(dict(zip(self.config["reduce_key"], key)))
+            fold_cost = response.total_cost
+
+            return folded_output, fold_prompt, fold_cost
+
+        return None, fold_prompt, fold_cost
+
+    def _merge_results(
+        self, key: Tuple, outputs: List[Dict]
+    ) -> Tuple[Optional[Dict], str, float]:
+        """
+        Merge multiple outputs into a single result.
+
+        This method merges a list of outputs using the merge prompt.
+
+        Args:
+            key (Tuple): The reduce key tuple for the group.
+            outputs (List[Dict]): The list of outputs to be merged.
+
+        Returns:
+            Tuple[Optional[Dict], str, float]: A tuple containing the merged output (or None if processing failed),
+            the prompt used, and the cost of the merge operation.
+        """
+        start_time = time.time()
+        merge_prompt_template = Template(self.config["merge_prompt"])
+        merge_prompt = merge_prompt_template.render(
+            outputs=outputs, reduce_key=dict(zip(self.config["reduce_key"], key))
+        )
+        response = self.runner.api.call_llm(
+            self.config.get("model", self.default_model),
+            "merge",
+            [{"role": "user", "content": merge_prompt}],
+            self.config["output"]["schema"],
+            timeout_seconds=self.config.get("timeout", 120),
+            max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+            validation_config=(
+                {
+                    "num_retries": self.num_retries_on_validate_failure,
+                    "val_rule": self.config.get("validate", []),
+                    "validation_fn": self.validation_fn,
+                }
+                if self.config.get("validate", None)
+                else None
+            ),
+            bypass_cache=self.config.get("bypass_cache", False),
+            verbose=self.config.get("verbose", False),
+            litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+        )
+
+        end_time = time.time()
+        self._update_merge_time(end_time - start_time)
+
+        if response.validated:
+            merged_output = self.runner.api.parse_llm_response(
+                response.response,
+                schema=self.config["output"]["schema"],
+                manually_fix_errors=self.manually_fix_errors,
+            )[0]
+            merged_output.update(dict(zip(self.config["reduce_key"], key)))
+            merge_cost = response.total_cost
+            return merged_output, merge_prompt, merge_cost
+
+        return None, merge_prompt, merge_cost
+
+    def get_fold_time(self) -> Tuple[float, bool]:
+        """
+        Get the average fold time or a default value.
+
+        Returns:
+            Tuple[float, bool]: A tuple containing the average fold time (or default) and a boolean
+            indicating whether the default value was used.
+        """
+        if "fold_time" in self.config:
+            return self.config["fold_time"], False
+        with self.lock:
+            if len(self.fold_times) >= self.min_samples:
+                return sum(self.fold_times) / len(self.fold_times), False
+        return 1.0, True  # Default to 1 second if no data is available
+
+    def get_merge_time(self) -> Tuple[float, bool]:
+        """
+        Get the average merge time or a default value.
+
+        Returns:
+            Tuple[float, bool]: A tuple containing the average merge time (or default) and a boolean
+            indicating whether the default value was used.
+        """
+        if "merge_time" in self.config:
+            return self.config["merge_time"], False
+        with self.lock:
+            if len(self.merge_times) >= self.min_samples:
+                return sum(self.merge_times) / len(self.merge_times), False
+        return 1.0, True  # Default to 1 second if no data is available
+
+    def _update_fold_time(self, time: float) -> None:
+        """
+        Update the fold time statistics.
+
+        Args:
+            time (float): The time taken for a fold operation.
+        """
+        with self.lock:
+            self.fold_times.append(time)
+
+    def _update_merge_time(self, time: float) -> None:
+        """
+        Update the merge time statistics.
+
+        Args:
+            time (float): The time taken for a merge operation.
+        """
+        with self.lock:
+            self.merge_times.append(time)
+
+    def _batch_reduce(
+        self, key: Tuple, group_list: List[Dict], scratchpad: Optional[str] = None
+    ) -> Tuple[Optional[Dict], str, float]:
+        """
+        Perform a batch reduce operation on a group of items.
+
+        This method reduces a group of items into a single output using the reduce prompt.
+
+        Args:
+            key (Tuple): The reduce key tuple for the group.
+            group_list (List[Dict]): The list of items to be reduced.
+            scratchpad (Optional[str]): The scratchpad to use for the reduce operation.
+        Returns:
+            Tuple[Optional[Dict], str, float]: A tuple containing the reduced output (or None if processing failed),
+            the prompt used, and the cost of the reduce operation.
+        """
+        prompt_template = Template(self.config["prompt"])
+        prompt = prompt_template.render(
+            reduce_key=dict(zip(self.config["reduce_key"], key)), inputs=group_list
+        )
+        item_cost = 0
+
+        response = self.runner.api.call_llm(
+            self.config.get("model", self.default_model),
+            "reduce",
+            [{"role": "user", "content": prompt}],
+            self.config["output"]["schema"],
+            scratchpad=scratchpad,
+            timeout_seconds=self.config.get("timeout", 120),
+            max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+            bypass_cache=self.config.get("bypass_cache", False),
+            validation_config=(
+                {
+                    "num_retries": self.num_retries_on_validate_failure,
+                    "val_rule": self.config.get("validate", []),
+                    "validation_fn": self.validation_fn,
+                }
+                if self.config.get("validate", None)
+                else None
+            ),
+            gleaning_config=self.config.get("gleaning", None),
+            verbose=self.config.get("verbose", False),
+            litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+        )
+
+        item_cost += response.total_cost
+
+        if response.validated:
+            output = self.runner.api.parse_llm_response(
+                response.response,
+                schema=self.config["output"]["schema"],
+                manually_fix_errors=self.manually_fix_errors,
+            )[0]
+            output.update(dict(zip(self.config["reduce_key"], key)))
+
+            return output, prompt, item_cost
+        return None, prompt, item_cost
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ __init__(*args, **kwargs) + +

+ + +
+ +

Initialize the ReduceOperation.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ *args + + +
+

Variable length argument list.

+
+
+ () +
+ **kwargs + + +
+

Arbitrary keyword arguments.

+
+
+ {} +
+ +
+ Source code in docetl/operations/reduce.py +
58
+59
+60
+61
+62
+63
+64
+65
+66
+67
+68
+69
+70
+71
+72
+73
+74
+75
+76
+77
+78
def __init__(self, *args, **kwargs):
+    """
+    Initialize the ReduceOperation.
+
+    Args:
+        *args: Variable length argument list.
+        **kwargs: Arbitrary keyword arguments.
+    """
+    super().__init__(*args, **kwargs)
+    self.min_samples = 5
+    self.max_samples = 1000
+    self.fold_times = deque(maxlen=self.max_samples)
+    self.merge_times = deque(maxlen=self.max_samples)
+    self.lock = Lock()
+    self.config["reduce_key"] = (
+        [self.config["reduce_key"]]
+        if isinstance(self.config["reduce_key"], str)
+        else self.config["reduce_key"]
+    )
+    self.intermediates = {}
+    self.lineage_keys = self.config.get("output", {}).get("lineage", [])
+
+
+
+ +
+ +
+ + +

+ execute(input_data) + +

+ + +
+ +

Execute the reduce operation on the provided input data.

+

This method sorts and groups the input data by the reduce key(s), then processes each group +using either parallel fold and merge, incremental reduce, or batch reduce strategies.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ input_data + + List[Dict] + +
+

The input data to process.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[List[Dict], float] + +
+

Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.

+
+
+ +
+ Source code in docetl/operations/reduce.py +
297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+    """
+    Execute the reduce operation on the provided input data.
+
+    This method sorts and groups the input data by the reduce key(s), then processes each group
+    using either parallel fold and merge, incremental reduce, or batch reduce strategies.
+
+    Args:
+        input_data (List[Dict]): The input data to process.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.
+    """
+    if self.config.get("gleaning", {}).get("validation_prompt", None):
+        self.console.log(
+            f"Using gleaning with validation prompt: {self.config.get('gleaning', {}).get('validation_prompt', '')}"
+        )
+
+    reduce_keys = self.config["reduce_key"]
+    if isinstance(reduce_keys, str):
+        reduce_keys = [reduce_keys]
+    input_schema = self.config.get("input", {}).get("schema", {})
+
+    if self.status:
+        self.status.stop()
+
+    # Check if we need to group everything into one group
+    if reduce_keys == ["_all"] or reduce_keys == "_all":
+        grouped_data = [("_all", input_data)]
+    else:
+        # Group the input data by the reduce key(s) while maintaining original order
+        def get_group_key(item):
+            key_values = []
+            for key in reduce_keys:
+                value = item[key]
+                # Special handling for list-type values
+                if isinstance(value, list):
+                    key_values.append(tuple(sorted(value)))  # Convert list to sorted tuple
+                else:
+                    key_values.append(value)
+            return tuple(key_values)
+
+        grouped_data = {}
+        for item in input_data:
+            key = get_group_key(item)
+            if key not in grouped_data:
+                grouped_data[key] = []
+            grouped_data[key].append(item)
+
+        # Convert the grouped data to a list of tuples
+        grouped_data = list(grouped_data.items())
+
+    def process_group(
+        key: Tuple, group_elems: List[Dict]
+    ) -> Tuple[Optional[Dict], float]:
+        if input_schema:
+            group_list = [
+                {k: item[k] for k in input_schema.keys() if k in item}
+                for item in group_elems
+            ]
+        else:
+            group_list = group_elems
+
+        total_cost = 0.0
+
+        # Apply value sampling if enabled
+        value_sampling = self.config.get("value_sampling", {})
+        if value_sampling.get("enabled", False):
+            sample_size = min(value_sampling["sample_size"], len(group_list))
+            method = value_sampling["method"]
+
+            if method == "random":
+                group_sample = random.sample(group_list, sample_size)
+                group_sample.sort(key=lambda x: group_list.index(x))
+            elif method == "first_n":
+                group_sample = group_list[:sample_size]
+            elif method == "cluster":
+                group_sample, embedding_cost = self._cluster_based_sampling(
+                    group_list, value_sampling, sample_size
+                )
+                group_sample.sort(key=lambda x: group_list.index(x))
+                total_cost += embedding_cost
+            elif method == "sem_sim":
+                group_sample, embedding_cost = self._semantic_similarity_sampling(
+                    key, group_list, value_sampling, sample_size
+                )
+                group_sample.sort(key=lambda x: group_list.index(x))
+                total_cost += embedding_cost
+
+            group_list = group_sample
+
+        # Only execute merge-based plans if associative = True
+        if "merge_prompt" in self.config and self.config.get("associative", True):
+            result, prompts, cost = self._parallel_fold_and_merge(key, group_list)
+        elif (
+            self.config.get("fold_batch_size", None)
+            and self.config.get("fold_batch_size") >= len(group_list)
+        ):
+            # If the fold batch size is greater than or equal to the number of items in the group,
+            # we can just run a single fold operation
+            result, prompt, cost = self._batch_reduce(key, group_list)
+            prompts = [prompt]
+        elif "fold_prompt" in self.config:
+            result, prompts, cost = self._incremental_reduce(key, group_list)
+        else:
+            result, prompt, cost = self._batch_reduce(key, group_list)
+            prompts = [prompt]
+
+        total_cost += cost
+
+        # Add the counts of items in the group to the result
+        result[f"_counts_prereduce_{self.config['name']}"] = len(group_elems)
+
+        if self.config.get("enable_observability", False):
+            # Add the _observability_{self.config['name']} key to the result
+            result[f"_observability_{self.config['name']}"] = {
+                "prompts": prompts
+            }
+
+        # Apply pass-through at the group level
+        if (
+            result is not None
+            and self.config.get("pass_through", False)
+            and group_elems
+        ):
+            for k, v in group_elems[0].items():
+                if k not in self.config["output"]["schema"] and k not in result:
+                    result[k] = v
+
+        # Add lineage information
+        if result is not None and self.lineage_keys:
+            lineage = []
+            for item in group_elems:
+                lineage_item = {
+                    k: item.get(k) for k in self.lineage_keys if k in item
+                }
+                if lineage_item:
+                    lineage.append(lineage_item)
+            result[f"{self.config['name']}_lineage"] = lineage
+
+        return result, total_cost
+
+    with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+        futures = [
+            executor.submit(process_group, key, group)
+            for key, group in grouped_data
+        ]
+        results = []
+        total_cost = 0
+        for future in rich_as_completed(
+            futures,
+            total=len(futures),
+            desc=f"Processing {self.config['name']} (reduce) on all documents",
+            leave=True,
+            console=self.console,
+        ):
+            output, item_cost = future.result()
+            total_cost += item_cost
+            if output is not None:
+                results.append(output)
+
+    if self.config.get("persist_intermediates", False):
+        for result in results:
+            key = tuple(result[k] for k in self.config["reduce_key"])
+            if key in self.intermediates:
+                result[f"_{self.config['name']}_intermediates"] = (
+                    self.intermediates[key]
+                )
+
+    if self.status:
+        self.status.start()
+
+    return results, total_cost
+
+
+
+ +
+ +
+ + +

+ get_fold_time() + +

+ + +
+ +

Get the average fold time or a default value.

+ + +

Returns:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ float + +
+

Tuple[float, bool]: A tuple containing the average fold time (or default) and a boolean

+
+
+ bool + +
+

indicating whether the default value was used.

+
+
+ +
+ Source code in docetl/operations/reduce.py +
900
+901
+902
+903
+904
+905
+906
+907
+908
+909
+910
+911
+912
+913
def get_fold_time(self) -> Tuple[float, bool]:
+    """
+    Get the average fold time or a default value.
+
+    Returns:
+        Tuple[float, bool]: A tuple containing the average fold time (or default) and a boolean
+        indicating whether the default value was used.
+    """
+    if "fold_time" in self.config:
+        return self.config["fold_time"], False
+    with self.lock:
+        if len(self.fold_times) >= self.min_samples:
+            return sum(self.fold_times) / len(self.fold_times), False
+    return 1.0, True  # Default to 1 second if no data is available
+
+
+
+ +
+ +
+ + +

+ get_merge_time() + +

+ + +
+ +

Get the average merge time or a default value.

+ + +

Returns:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ float + +
+

Tuple[float, bool]: A tuple containing the average merge time (or default) and a boolean

+
+
+ bool + +
+

indicating whether the default value was used.

+
+
+ +
+ Source code in docetl/operations/reduce.py +
915
+916
+917
+918
+919
+920
+921
+922
+923
+924
+925
+926
+927
+928
def get_merge_time(self) -> Tuple[float, bool]:
+    """
+    Get the average merge time or a default value.
+
+    Returns:
+        Tuple[float, bool]: A tuple containing the average merge time (or default) and a boolean
+        indicating whether the default value was used.
+    """
+    if "merge_time" in self.config:
+        return self.config["merge_time"], False
+    with self.lock:
+        if len(self.merge_times) >= self.min_samples:
+            return sum(self.merge_times) / len(self.merge_times), False
+    return 1.0, True  # Default to 1 second if no data is available
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Perform comprehensive syntax checks on the configuration of the ReduceOperation.

+

This method validates the presence and correctness of all required configuration keys, Jinja2 templates, and ensures the correct +structure and types of the entire configuration.

+

The method performs the following checks: +1. Verifies the presence of all required keys in the configuration. +2. Validates the structure and content of the 'output' configuration, including its 'schema'. +3. Checks if the main 'prompt' is a valid Jinja2 template and contains the required 'inputs' variable. +4. If 'merge_prompt' is specified, ensures that 'fold_prompt' is also present. +5. If 'fold_prompt' is present, verifies the existence of 'fold_batch_size'. +6. Validates the 'fold_prompt' as a Jinja2 template with required variables 'inputs' and 'output'. +7. If present, checks 'merge_prompt' as a valid Jinja2 template with required 'outputs' variable. +8. Verifies types of various configuration inputs (e.g., 'fold_batch_size' as int). +9. Checks for the presence and validity of optional configurations like 'model'.

+ + +

Raises:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If any required configuration is missing, if templates are invalid or missing required + variables, or if any other configuration aspect is incorrect or inconsistent.

+
+
+ TypeError + +
+

If any configuration value has an incorrect type, such as 'schema' not being a dict + or 'fold_batch_size' not being an integer.

+
+
+ +
+ Source code in docetl/operations/reduce.py +
 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
def syntax_check(self) -> None:
+    """
+    Perform comprehensive syntax checks on the configuration of the ReduceOperation.
+
+    This method validates the presence and correctness of all required configuration keys, Jinja2 templates, and ensures the correct
+    structure and types of the entire configuration.
+
+    The method performs the following checks:
+    1. Verifies the presence of all required keys in the configuration.
+    2. Validates the structure and content of the 'output' configuration, including its 'schema'.
+    3. Checks if the main 'prompt' is a valid Jinja2 template and contains the required 'inputs' variable.
+    4. If 'merge_prompt' is specified, ensures that 'fold_prompt' is also present.
+    5. If 'fold_prompt' is present, verifies the existence of 'fold_batch_size'.
+    6. Validates the 'fold_prompt' as a Jinja2 template with required variables 'inputs' and 'output'.
+    7. If present, checks 'merge_prompt' as a valid Jinja2 template with required 'outputs' variable.
+    8. Verifies types of various configuration inputs (e.g., 'fold_batch_size' as int).
+    9. Checks for the presence and validity of optional configurations like 'model'.
+
+    Raises:
+        ValueError: If any required configuration is missing, if templates are invalid or missing required
+                    variables, or if any other configuration aspect is incorrect or inconsistent.
+        TypeError: If any configuration value has an incorrect type, such as 'schema' not being a dict
+                   or 'fold_batch_size' not being an integer.
+    """
+    required_keys = ["reduce_key", "prompt", "output"]
+    for key in required_keys:
+        if key not in self.config:
+            raise ValueError(
+                f"Missing required key '{key}' in {self.config['name']} configuration"
+            )
+
+    if "schema" not in self.config["output"]:
+        raise ValueError(
+            f"Missing 'schema' in {self.config['name']} 'output' configuration"
+        )
+
+    if not isinstance(self.config["output"]["schema"], dict):
+        raise TypeError(
+            f"'schema' in {self.config['name']} 'output' configuration must be a dictionary"
+        )
+
+    if not self.config["output"]["schema"]:
+        raise ValueError(
+            f"'schema' in {self.config['name']} 'output' configuration cannot be empty"
+        )
+
+    # Check if the prompt is a valid Jinja2 template
+    try:
+        template = Template(self.config["prompt"])
+        template_vars = template.environment.parse(self.config["prompt"]).find_all(
+            jinja2.nodes.Name
+        )
+        template_var_names = {var.name for var in template_vars}
+        if "inputs" not in template_var_names:
+            raise ValueError(
+                f"Prompt template for {self.config['name']} must include the 'inputs' variable"
+            )
+    except Exception as e:
+        raise ValueError(
+            f"Invalid Jinja2 template in {self.config['name']} 'prompt': {str(e)}"
+        )
+
+    # Check if fold_prompt is a valid Jinja2 template (now required if merge exists)
+    if "merge_prompt" in self.config:
+        if "fold_prompt" not in self.config:
+            raise ValueError(
+                f"'fold_prompt' is required when 'merge_prompt' is specified in {self.config['name']}"
+            )
+
+    if "fold_prompt" in self.config:
+        if "fold_batch_size" not in self.config:
+            raise ValueError(
+                f"'fold_batch_size' is required when 'fold_prompt' is specified in {self.config['name']}"
+            )
+
+        try:
+            fold_template = Template(self.config["fold_prompt"])
+            fold_template_vars = fold_template.environment.parse(
+                self.config["fold_prompt"]
+            ).find_all(jinja2.nodes.Name)
+            fold_template_var_names = {var.name for var in fold_template_vars}
+            required_vars = {"inputs", "output"}
+            if not required_vars.issubset(fold_template_var_names):
+                raise ValueError(
+                    f"Fold template in {self.config['name']} must include variables: {required_vars}. Current template includes: {fold_template_var_names}"
+                )
+        except Exception as e:
+            raise ValueError(
+                f"Invalid Jinja2 template in {self.config['name']} 'fold_prompt': {str(e)}"
+            )
+
+    # Check merge_prompt and merge_batch_size
+    if "merge_prompt" in self.config:
+        if "merge_batch_size" not in self.config:
+            raise ValueError(
+                f"'merge_batch_size' is required when 'merge_prompt' is specified in {self.config['name']}"
+            )
+
+        try:
+            merge_template = Template(self.config["merge_prompt"])
+            merge_template_vars = merge_template.environment.parse(
+                self.config["merge_prompt"]
+            ).find_all(jinja2.nodes.Name)
+            merge_template_var_names = {var.name for var in merge_template_vars}
+            if "outputs" not in merge_template_var_names:
+                raise ValueError(
+                    f"Merge template in {self.config['name']} must include the 'outputs' variable"
+                )
+        except Exception as e:
+            raise ValueError(
+                f"Invalid Jinja2 template in {self.config['name']} 'merge_prompt': {str(e)}"
+            )
+
+    # Check if the model is specified (optional)
+    if "model" in self.config and not isinstance(self.config["model"], str):
+        raise TypeError(
+            f"'model' in {self.config['name']} configuration must be a string"
+        )
+
+    # Check if reduce_key is a string or a list of strings
+    if not isinstance(self.config["reduce_key"], (str, list)):
+        raise TypeError(
+            f"'reduce_key' in {self.config['name']} configuration must be a string or a list of strings"
+        )
+    if isinstance(self.config["reduce_key"], list):
+        if not all(isinstance(key, str) for key in self.config["reduce_key"]):
+            raise TypeError(
+                f"All elements in 'reduce_key' list in {self.config['name']} configuration must be strings"
+            )
+
+    # Check if input schema is provided and valid (optional)
+    if "input" in self.config:
+        if "schema" not in self.config["input"]:
+            raise ValueError(
+                f"Missing 'schema' in {self.config['name']} 'input' configuration"
+            )
+        if not isinstance(self.config["input"]["schema"], dict):
+            raise TypeError(
+                f"'schema' in {self.config['name']} 'input' configuration must be a dictionary"
+            )
+
+    # Check if fold_batch_size and merge_batch_size are positive integers
+    for key in ["fold_batch_size", "merge_batch_size"]:
+        if key in self.config:
+            if not isinstance(self.config[key], int) or self.config[key] <= 0:
+                raise ValueError(
+                    f"'{key}' in {self.config['name']} configuration must be a positive integer"
+                )
+
+    if "value_sampling" in self.config:
+        sampling = self.config["value_sampling"]
+        if not isinstance(sampling, dict):
+            raise TypeError(
+                f"'value_sampling' in {self.config['name']} configuration must be a dictionary"
+            )
+
+        if "enabled" not in sampling:
+            raise ValueError(
+                f"'enabled' is required in {self.config['name']} 'value_sampling' configuration"
+            )
+        if not isinstance(sampling["enabled"], bool):
+            raise TypeError(
+                f"'enabled' in {self.config['name']} 'value_sampling' configuration must be a boolean"
+            )
+
+        if sampling["enabled"]:
+            if "sample_size" not in sampling:
+                raise ValueError(
+                    f"'sample_size' is required when value_sampling is enabled in {self.config['name']}"
+                )
+            if (
+                not isinstance(sampling["sample_size"], int)
+                or sampling["sample_size"] <= 0
+            ):
+                raise ValueError(
+                    f"'sample_size' in {self.config['name']} configuration must be a positive integer"
+                )
+
+            if "method" not in sampling:
+                raise ValueError(
+                    f"'method' is required when value_sampling is enabled in {self.config['name']}"
+                )
+            if sampling["method"] not in [
+                "random",
+                "first_n",
+                "cluster",
+                "sem_sim",
+            ]:
+                raise ValueError(
+                    f"Invalid 'method'. Must be 'random', 'first_n', or 'embedding' in {self.config['name']}"
+                )
+
+            if sampling["method"] == "embedding":
+                if "embedding_model" not in sampling:
+                    raise ValueError(
+                        f"'embedding_model' is required when using embedding-based sampling in {self.config['name']}"
+                    )
+                if "embedding_keys" not in sampling:
+                    raise ValueError(
+                        f"'embedding_keys' is required when using embedding-based sampling in {self.config['name']}"
+                    )
+
+    # Check if lineage is a list of strings
+    if "lineage" in self.config.get("output", {}):
+        if not isinstance(self.config["output"]["lineage"], list):
+            raise TypeError(
+                f"'lineage' in {self.config['name']} 'output' configuration must be a list"
+            )
+        if not all(
+            isinstance(key, str) for key in self.config["output"]["lineage"]
+        ):
+            raise TypeError(
+                f"All elements in 'lineage' list in {self.config['name']} 'output' configuration must be strings"
+            )
+
+    self.gleaning_check()
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.operations.map.ParallelMapOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + + + + + + +
+ Source code in docetl/operations/map.py +
319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
+470
+471
+472
+473
+474
+475
+476
+477
+478
+479
+480
+481
+482
+483
+484
+485
+486
+487
+488
+489
+490
+491
+492
+493
+494
+495
+496
+497
+498
+499
+500
+501
+502
+503
+504
+505
+506
+507
+508
+509
+510
+511
+512
+513
+514
+515
+516
+517
+518
+519
+520
+521
+522
+523
+524
+525
+526
+527
+528
+529
+530
+531
+532
+533
class ParallelMapOperation(BaseOperation):
+    class schema(BaseOperation.schema):
+        type: str = "parallel_map"
+        prompts: List[Dict[str, Any]]
+        output: Dict[str, Any]
+        enable_observability: bool = False
+
+    def __init__(
+        self,
+        *args,
+        **kwargs,
+    ):
+        super().__init__(*args, **kwargs)
+
+    def syntax_check(self) -> None:
+        """
+        Checks the configuration of the ParallelMapOperation for required keys and valid structure.
+
+        Raises:
+            ValueError: If required keys are missing or if the configuration structure is invalid.
+            TypeError: If the configuration values have incorrect types.
+        """
+        if "drop_keys" in self.config:
+            if not isinstance(self.config["drop_keys"], list):
+                raise TypeError(
+                    "'drop_keys' in configuration must be a list of strings"
+                )
+            for key in self.config["drop_keys"]:
+                if not isinstance(key, str):
+                    raise TypeError("All items in 'drop_keys' must be strings")
+        elif "prompts" not in self.config:
+            raise ValueError(
+                "If 'drop_keys' is not specified, 'prompts' must be present in the configuration"
+            )
+
+        if "prompts" in self.config:
+            if not isinstance(self.config["prompts"], list):
+                raise ValueError(
+                    "ParallelMapOperation requires a 'prompts' list in the configuration"
+                )
+
+            if not self.config["prompts"]:
+                raise ValueError("The 'prompts' list cannot be empty")
+
+            for i, prompt_config in enumerate(self.config["prompts"]):
+                if not isinstance(prompt_config, dict):
+                    raise TypeError(f"Prompt configuration {i} must be a dictionary")
+
+                required_keys = ["prompt", "output_keys"]
+                for key in required_keys:
+                    if key not in prompt_config:
+                        raise ValueError(
+                            f"Missing required key '{key}' in prompt configuration {i}"
+                        )
+                if not isinstance(prompt_config["prompt"], str):
+                    raise TypeError(
+                        f"'prompt' in prompt configuration {i} must be a string"
+                    )
+
+                if not isinstance(prompt_config["output_keys"], list):
+                    raise TypeError(
+                        f"'output_keys' in prompt configuration {i} must be a list"
+                    )
+
+                if not prompt_config["output_keys"]:
+                    raise ValueError(
+                        f"'output_keys' list in prompt configuration {i} cannot be empty"
+                    )
+
+                # Check if the prompt is a valid Jinja2 template
+                try:
+                    Template(prompt_config["prompt"])
+                except Exception as e:
+                    raise ValueError(
+                        f"Invalid Jinja2 template in prompt configuration {i}: {str(e)}"
+                    ) from e
+
+                # Check if the model is specified (optional)
+                if "model" in prompt_config and not isinstance(
+                    prompt_config["model"], str
+                ):
+                    raise TypeError(
+                        f"'model' in prompt configuration {i} must be a string"
+                    )
+
+            # Check if all output schema keys are covered by the prompts
+            output_schema = self.config["output"]["schema"]
+            output_keys_covered = set()
+            for prompt_config in self.config["prompts"]:
+                output_keys_covered.update(prompt_config["output_keys"])
+
+            missing_keys = set(output_schema.keys()) - output_keys_covered
+            if missing_keys:
+                raise ValueError(
+                    f"The following output schema keys are not covered by any prompt: {missing_keys}"
+                )
+
+    def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+        """
+        Executes the parallel map operation on the provided input data.
+
+        Args:
+            input_data (List[Dict]): The input data to process.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.
+
+        This method performs the following steps:
+        1. If prompts are specified, it processes each input item using multiple prompts in parallel
+        2. Aggregates results from different prompts for each input item
+        3. Validates the combined output for each item
+        4. If drop_keys is specified, it drops the specified keys from each document
+        5. Calculates total cost of the operation
+        """
+        results = {}
+        total_cost = 0
+        output_schema = self.config.get("output", {}).get("schema", {})
+
+        # Check if there's no prompt and only drop_keys
+        if "prompts" not in self.config and "drop_keys" in self.config:
+            # If only drop_keys is specified, simply drop the keys and return
+            dropped_results = []
+            for item in input_data:
+                new_item = {
+                    k: v for k, v in item.items() if k not in self.config["drop_keys"]
+                }
+                dropped_results.append(new_item)
+            return dropped_results, 0.0  # Return the modified data with no cost
+
+        if self.status:
+            self.status.stop()
+
+        def process_prompt(item, prompt_config):
+            prompt = render_jinja_template(prompt_config["prompt"], item)
+            local_output_schema = {
+                key: output_schema[key] for key in prompt_config["output_keys"]
+            }
+            model = prompt_config.get("model", self.default_model)
+            if not model:
+                model = self.default_model
+
+            # Start of Selection
+            # If there are tools, we need to pass in the tools
+            response = self.runner.api.call_llm(
+                model,
+                "parallel_map",
+                [{"role": "user", "content": prompt}],
+                local_output_schema,
+                tools=prompt_config.get("tools", None),
+                timeout_seconds=self.config.get("timeout", 120),
+                max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+                bypass_cache=self.config.get("bypass_cache", False),
+                litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+            )
+            output = self.runner.api.parse_llm_response(
+                response.response,
+                schema=local_output_schema,
+                tools=prompt_config.get("tools", None),
+                manually_fix_errors=self.manually_fix_errors,
+            )[0]
+            return output, prompt, response.total_cost
+
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            if "prompts" in self.config:
+                # Create all futures at once
+                all_futures = [
+                    executor.submit(process_prompt, item, prompt_config)
+                    for item in input_data
+                    for prompt_config in self.config["prompts"]
+                ]
+
+                # Process results in order
+                for i in tqdm(
+                    range(len(all_futures)),
+                    desc="Processing parallel map items",
+                ):
+                    future = all_futures[i]
+                    output, prompt, cost = future.result()
+                    total_cost += cost
+
+                    # Determine which item this future corresponds to
+                    item_index = i // len(self.config["prompts"])
+                    prompt_index = i % len(self.config["prompts"])
+
+                    # Initialize or update the item_result
+                    if prompt_index == 0:
+                        item_result = input_data[item_index].copy()
+                        results[item_index] = item_result
+
+                    # Fetch the item_result
+                    item_result = results[item_index]
+
+                    if self.config.get("enable_observability", False):
+                        if f"_observability_{self.config['name']}" not in item_result:
+                            item_result[f"_observability_{self.config['name']}"] = {}
+                        item_result[f"_observability_{self.config['name']}"].update({f"prompt_{prompt_index}": prompt})
+
+                    # Update the item_result with the output
+                    item_result.update(output)
+
+            else:
+                results = {i: item.copy() for i, item in enumerate(input_data)}
+
+        # Apply drop_keys if specified
+        if "drop_keys" in self.config:
+            drop_keys = self.config["drop_keys"]
+            for item in results.values():
+                for key in drop_keys:
+                    item.pop(key, None)
+
+        if self.status:
+            self.status.start()
+
+        # Return the results in order
+        return [results[i] for i in range(len(input_data)) if i in results], total_cost
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ execute(input_data) + +

+ + +
+ +

Executes the parallel map operation on the provided input data.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ input_data + + List[Dict] + +
+

The input data to process.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[List[Dict], float] + +
+

Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.

+
+
+

This method performs the following steps: +1. If prompts are specified, it processes each input item using multiple prompts in parallel +2. Aggregates results from different prompts for each input item +3. Validates the combined output for each item +4. If drop_keys is specified, it drops the specified keys from each document +5. Calculates total cost of the operation

+ +
+ Source code in docetl/operations/map.py +
416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
+470
+471
+472
+473
+474
+475
+476
+477
+478
+479
+480
+481
+482
+483
+484
+485
+486
+487
+488
+489
+490
+491
+492
+493
+494
+495
+496
+497
+498
+499
+500
+501
+502
+503
+504
+505
+506
+507
+508
+509
+510
+511
+512
+513
+514
+515
+516
+517
+518
+519
+520
+521
+522
+523
+524
+525
+526
+527
+528
+529
+530
+531
+532
+533
def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+    """
+    Executes the parallel map operation on the provided input data.
+
+    Args:
+        input_data (List[Dict]): The input data to process.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the processed results and the total cost of the operation.
+
+    This method performs the following steps:
+    1. If prompts are specified, it processes each input item using multiple prompts in parallel
+    2. Aggregates results from different prompts for each input item
+    3. Validates the combined output for each item
+    4. If drop_keys is specified, it drops the specified keys from each document
+    5. Calculates total cost of the operation
+    """
+    results = {}
+    total_cost = 0
+    output_schema = self.config.get("output", {}).get("schema", {})
+
+    # Check if there's no prompt and only drop_keys
+    if "prompts" not in self.config and "drop_keys" in self.config:
+        # If only drop_keys is specified, simply drop the keys and return
+        dropped_results = []
+        for item in input_data:
+            new_item = {
+                k: v for k, v in item.items() if k not in self.config["drop_keys"]
+            }
+            dropped_results.append(new_item)
+        return dropped_results, 0.0  # Return the modified data with no cost
+
+    if self.status:
+        self.status.stop()
+
+    def process_prompt(item, prompt_config):
+        prompt = render_jinja_template(prompt_config["prompt"], item)
+        local_output_schema = {
+            key: output_schema[key] for key in prompt_config["output_keys"]
+        }
+        model = prompt_config.get("model", self.default_model)
+        if not model:
+            model = self.default_model
+
+        # Start of Selection
+        # If there are tools, we need to pass in the tools
+        response = self.runner.api.call_llm(
+            model,
+            "parallel_map",
+            [{"role": "user", "content": prompt}],
+            local_output_schema,
+            tools=prompt_config.get("tools", None),
+            timeout_seconds=self.config.get("timeout", 120),
+            max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+            bypass_cache=self.config.get("bypass_cache", False),
+            litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+        )
+        output = self.runner.api.parse_llm_response(
+            response.response,
+            schema=local_output_schema,
+            tools=prompt_config.get("tools", None),
+            manually_fix_errors=self.manually_fix_errors,
+        )[0]
+        return output, prompt, response.total_cost
+
+    with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+        if "prompts" in self.config:
+            # Create all futures at once
+            all_futures = [
+                executor.submit(process_prompt, item, prompt_config)
+                for item in input_data
+                for prompt_config in self.config["prompts"]
+            ]
+
+            # Process results in order
+            for i in tqdm(
+                range(len(all_futures)),
+                desc="Processing parallel map items",
+            ):
+                future = all_futures[i]
+                output, prompt, cost = future.result()
+                total_cost += cost
+
+                # Determine which item this future corresponds to
+                item_index = i // len(self.config["prompts"])
+                prompt_index = i % len(self.config["prompts"])
+
+                # Initialize or update the item_result
+                if prompt_index == 0:
+                    item_result = input_data[item_index].copy()
+                    results[item_index] = item_result
+
+                # Fetch the item_result
+                item_result = results[item_index]
+
+                if self.config.get("enable_observability", False):
+                    if f"_observability_{self.config['name']}" not in item_result:
+                        item_result[f"_observability_{self.config['name']}"] = {}
+                    item_result[f"_observability_{self.config['name']}"].update({f"prompt_{prompt_index}": prompt})
+
+                # Update the item_result with the output
+                item_result.update(output)
+
+        else:
+            results = {i: item.copy() for i, item in enumerate(input_data)}
+
+    # Apply drop_keys if specified
+    if "drop_keys" in self.config:
+        drop_keys = self.config["drop_keys"]
+        for item in results.values():
+            for key in drop_keys:
+                item.pop(key, None)
+
+    if self.status:
+        self.status.start()
+
+    # Return the results in order
+    return [results[i] for i in range(len(input_data)) if i in results], total_cost
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Checks the configuration of the ParallelMapOperation for required keys and valid structure.

+ + +

Raises:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If required keys are missing or if the configuration structure is invalid.

+
+
+ TypeError + +
+

If the configuration values have incorrect types.

+
+
+ +
+ Source code in docetl/operations/map.py +
333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
def syntax_check(self) -> None:
+    """
+    Checks the configuration of the ParallelMapOperation for required keys and valid structure.
+
+    Raises:
+        ValueError: If required keys are missing or if the configuration structure is invalid.
+        TypeError: If the configuration values have incorrect types.
+    """
+    if "drop_keys" in self.config:
+        if not isinstance(self.config["drop_keys"], list):
+            raise TypeError(
+                "'drop_keys' in configuration must be a list of strings"
+            )
+        for key in self.config["drop_keys"]:
+            if not isinstance(key, str):
+                raise TypeError("All items in 'drop_keys' must be strings")
+    elif "prompts" not in self.config:
+        raise ValueError(
+            "If 'drop_keys' is not specified, 'prompts' must be present in the configuration"
+        )
+
+    if "prompts" in self.config:
+        if not isinstance(self.config["prompts"], list):
+            raise ValueError(
+                "ParallelMapOperation requires a 'prompts' list in the configuration"
+            )
+
+        if not self.config["prompts"]:
+            raise ValueError("The 'prompts' list cannot be empty")
+
+        for i, prompt_config in enumerate(self.config["prompts"]):
+            if not isinstance(prompt_config, dict):
+                raise TypeError(f"Prompt configuration {i} must be a dictionary")
+
+            required_keys = ["prompt", "output_keys"]
+            for key in required_keys:
+                if key not in prompt_config:
+                    raise ValueError(
+                        f"Missing required key '{key}' in prompt configuration {i}"
+                    )
+            if not isinstance(prompt_config["prompt"], str):
+                raise TypeError(
+                    f"'prompt' in prompt configuration {i} must be a string"
+                )
+
+            if not isinstance(prompt_config["output_keys"], list):
+                raise TypeError(
+                    f"'output_keys' in prompt configuration {i} must be a list"
+                )
+
+            if not prompt_config["output_keys"]:
+                raise ValueError(
+                    f"'output_keys' list in prompt configuration {i} cannot be empty"
+                )
+
+            # Check if the prompt is a valid Jinja2 template
+            try:
+                Template(prompt_config["prompt"])
+            except Exception as e:
+                raise ValueError(
+                    f"Invalid Jinja2 template in prompt configuration {i}: {str(e)}"
+                ) from e
+
+            # Check if the model is specified (optional)
+            if "model" in prompt_config and not isinstance(
+                prompt_config["model"], str
+            ):
+                raise TypeError(
+                    f"'model' in prompt configuration {i} must be a string"
+                )
+
+        # Check if all output schema keys are covered by the prompts
+        output_schema = self.config["output"]["schema"]
+        output_keys_covered = set()
+        for prompt_config in self.config["prompts"]:
+            output_keys_covered.update(prompt_config["output_keys"])
+
+        missing_keys = set(output_schema.keys()) - output_keys_covered
+        if missing_keys:
+            raise ValueError(
+                f"The following output schema keys are not covered by any prompt: {missing_keys}"
+            )
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.operations.filter.FilterOperation + + +

+ + +
+

+ Bases: MapOperation

+ + + + + + + +
+ Source code in docetl/operations/filter.py +
 11
+ 12
+ 13
+ 14
+ 15
+ 16
+ 17
+ 18
+ 19
+ 20
+ 21
+ 22
+ 23
+ 24
+ 25
+ 26
+ 27
+ 28
+ 29
+ 30
+ 31
+ 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
class FilterOperation(MapOperation):
+    class schema(MapOperation.schema):
+        type: str = "filter"
+
+    def syntax_check(self) -> None:
+        """
+        Checks the configuration of the FilterOperation for required keys and valid structure.
+
+        Raises:
+            ValueError: If required keys are missing or if the output schema structure is invalid.
+            TypeError: If the schema in the output configuration is not a dictionary or if the schema value is not of type bool.
+
+        This method checks for the following:
+        - Presence of required keys: 'prompt' and 'output'
+        - Presence of 'schema' in the 'output' configuration
+        - The 'schema' is a non-empty dictionary with exactly one key-value pair
+        - The value in the schema is of type bool
+        """
+        required_keys = ["prompt", "output"]
+        for key in required_keys:
+            if key not in self.config:
+                raise ValueError(
+                    f"Missing required key '{key}' in FilterOperation configuration"
+                )
+
+        if "schema" not in self.config["output"]:
+            raise ValueError("Missing 'schema' in 'output' configuration")
+
+        if not isinstance(self.config["output"]["schema"], dict):
+            raise TypeError("'schema' in 'output' configuration must be a dictionary")
+
+        if not self.config["output"]["schema"]:
+            raise ValueError("'schema' in 'output' configuration cannot be empty")
+
+        schema = self.config["output"]["schema"]
+        if "_short_explanation" in schema:
+            schema = {k: v for k, v in schema.items() if k != "_short_explanation"}
+        if len(schema) != 1:
+            raise ValueError(
+                "The 'schema' in 'output' configuration must have exactly one key-value pair that maps to a boolean value"
+            )
+
+        key, value = next(iter(schema.items()))
+        if value not in ["bool", "boolean"]:
+            raise TypeError(
+                f"The value in the 'schema' must be of type bool, got {value}"
+            )
+
+    def execute(
+        self, input_data: List[Dict], is_build: bool = False
+    ) -> Tuple[List[Dict], float]:
+        """
+        Executes the filter operation on the input data.
+
+        Args:
+            input_data (List[Dict]): A list of dictionaries to process.
+            is_build (bool): Whether the operation is being executed in the build phase. Defaults to False.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the filtered list of dictionaries
+            and the total cost of the operation.
+
+        This method performs the following steps:
+        1. Processes each input item using an LLM model
+        2. Validates the output
+        3. Filters the results based on the specified filter key
+        4. Calculates the total cost of the operation
+
+        The method uses multi-threading to process items in parallel, improving performance
+        for large datasets.
+
+        Usage:
+        ```python
+        from docetl.operations import FilterOperation
+
+        config = {
+            "prompt": "Determine if the following item is important: {{input}}",
+            "output": {
+                "schema": {"is_important": "bool"}
+            },
+            "model": "gpt-3.5-turbo"
+        }
+        filter_op = FilterOperation(config)
+        input_data = [
+            {"id": 1, "text": "Critical update"},
+            {"id": 2, "text": "Regular maintenance"}
+        ]
+        results, cost = filter_op.execute(input_data)
+        print(f"Filtered results: {results}")
+        print(f"Total cost: {cost}")
+        ```
+        """
+        filter_key = next(
+            iter(
+                [
+                    k
+                    for k in self.config["output"]["schema"].keys()
+                    if k != "_short_explanation"
+                ]
+            )
+        )
+
+        results, total_cost = super().execute(input_data)
+
+        # Drop records with filter_key values that are False
+        results = [result for result in results if result[filter_key]]
+
+        return results, total_cost
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ execute(input_data, is_build=False) + +

+ + +
+ +

Executes the filter operation on the input data.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ input_data + + List[Dict] + +
+

A list of dictionaries to process.

+
+
+ required +
+ is_build + + bool + +
+

Whether the operation is being executed in the build phase. Defaults to False.

+
+
+ False +
+ + +

Returns:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ List[Dict] + +
+

Tuple[List[Dict], float]: A tuple containing the filtered list of dictionaries

+
+
+ float + +
+

and the total cost of the operation.

+
+
+

This method performs the following steps: +1. Processes each input item using an LLM model +2. Validates the output +3. Filters the results based on the specified filter key +4. Calculates the total cost of the operation

+

The method uses multi-threading to process items in parallel, improving performance +for large datasets.

+

Usage: +

from docetl.operations import FilterOperation
+
+config = {
+    "prompt": "Determine if the following item is important: {{input}}",
+    "output": {
+        "schema": {"is_important": "bool"}
+    },
+    "model": "gpt-3.5-turbo"
+}
+filter_op = FilterOperation(config)
+input_data = [
+    {"id": 1, "text": "Critical update"},
+    {"id": 2, "text": "Regular maintenance"}
+]
+results, cost = filter_op.execute(input_data)
+print(f"Filtered results: {results}")
+print(f"Total cost: {cost}")
+

+ +
+ Source code in docetl/operations/filter.py +
 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
def execute(
+    self, input_data: List[Dict], is_build: bool = False
+) -> Tuple[List[Dict], float]:
+    """
+    Executes the filter operation on the input data.
+
+    Args:
+        input_data (List[Dict]): A list of dictionaries to process.
+        is_build (bool): Whether the operation is being executed in the build phase. Defaults to False.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the filtered list of dictionaries
+        and the total cost of the operation.
+
+    This method performs the following steps:
+    1. Processes each input item using an LLM model
+    2. Validates the output
+    3. Filters the results based on the specified filter key
+    4. Calculates the total cost of the operation
+
+    The method uses multi-threading to process items in parallel, improving performance
+    for large datasets.
+
+    Usage:
+    ```python
+    from docetl.operations import FilterOperation
+
+    config = {
+        "prompt": "Determine if the following item is important: {{input}}",
+        "output": {
+            "schema": {"is_important": "bool"}
+        },
+        "model": "gpt-3.5-turbo"
+    }
+    filter_op = FilterOperation(config)
+    input_data = [
+        {"id": 1, "text": "Critical update"},
+        {"id": 2, "text": "Regular maintenance"}
+    ]
+    results, cost = filter_op.execute(input_data)
+    print(f"Filtered results: {results}")
+    print(f"Total cost: {cost}")
+    ```
+    """
+    filter_key = next(
+        iter(
+            [
+                k
+                for k in self.config["output"]["schema"].keys()
+                if k != "_short_explanation"
+            ]
+        )
+    )
+
+    results, total_cost = super().execute(input_data)
+
+    # Drop records with filter_key values that are False
+    results = [result for result in results if result[filter_key]]
+
+    return results, total_cost
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Checks the configuration of the FilterOperation for required keys and valid structure.

+ + +

Raises:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If required keys are missing or if the output schema structure is invalid.

+
+
+ TypeError + +
+

If the schema in the output configuration is not a dictionary or if the schema value is not of type bool.

+
+
+

This method checks for the following: +- Presence of required keys: 'prompt' and 'output' +- Presence of 'schema' in the 'output' configuration +- The 'schema' is a non-empty dictionary with exactly one key-value pair +- The value in the schema is of type bool

+ +
+ Source code in docetl/operations/filter.py +
15
+16
+17
+18
+19
+20
+21
+22
+23
+24
+25
+26
+27
+28
+29
+30
+31
+32
+33
+34
+35
+36
+37
+38
+39
+40
+41
+42
+43
+44
+45
+46
+47
+48
+49
+50
+51
+52
+53
+54
+55
+56
+57
def syntax_check(self) -> None:
+    """
+    Checks the configuration of the FilterOperation for required keys and valid structure.
+
+    Raises:
+        ValueError: If required keys are missing or if the output schema structure is invalid.
+        TypeError: If the schema in the output configuration is not a dictionary or if the schema value is not of type bool.
+
+    This method checks for the following:
+    - Presence of required keys: 'prompt' and 'output'
+    - Presence of 'schema' in the 'output' configuration
+    - The 'schema' is a non-empty dictionary with exactly one key-value pair
+    - The value in the schema is of type bool
+    """
+    required_keys = ["prompt", "output"]
+    for key in required_keys:
+        if key not in self.config:
+            raise ValueError(
+                f"Missing required key '{key}' in FilterOperation configuration"
+            )
+
+    if "schema" not in self.config["output"]:
+        raise ValueError("Missing 'schema' in 'output' configuration")
+
+    if not isinstance(self.config["output"]["schema"], dict):
+        raise TypeError("'schema' in 'output' configuration must be a dictionary")
+
+    if not self.config["output"]["schema"]:
+        raise ValueError("'schema' in 'output' configuration cannot be empty")
+
+    schema = self.config["output"]["schema"]
+    if "_short_explanation" in schema:
+        schema = {k: v for k, v in schema.items() if k != "_short_explanation"}
+    if len(schema) != 1:
+        raise ValueError(
+            "The 'schema' in 'output' configuration must have exactly one key-value pair that maps to a boolean value"
+        )
+
+    key, value = next(iter(schema.items()))
+    if value not in ["bool", "boolean"]:
+        raise TypeError(
+            f"The value in the 'schema' must be of type bool, got {value}"
+        )
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.operations.equijoin.EquijoinOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + + + + + + +
+ Source code in docetl/operations/equijoin.py +
 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
+470
+471
+472
+473
class EquijoinOperation(BaseOperation):
+    class schema(BaseOperation.schema):
+        type: str = "equijoin"
+        left: str
+        right: str
+        comparison_prompt: str
+        output: Optional[Dict[str, Any]] = None
+        blocking_threshold: Optional[float] = None
+        blocking_conditions: Optional[Dict[str, List[str]]] = None
+        limits: Optional[Dict[str, int]] = None
+        comparison_model: Optional[str] = None
+        optimize: Optional[bool] = None
+        embedding_model: Optional[str] = None
+        embedding_batch_size: Optional[int] = None
+        compare_batch_size: Optional[int] = None
+        limit_comparisons: Optional[int] = None
+        blocking_keys: Optional[Dict[str, List[str]]] = None
+        timeout: Optional[int] = None
+        litellm_completion_kwargs: Dict[str, Any] = Field(default_factory=dict)
+
+    def compare_pair(
+        self,
+        comparison_prompt: str,
+        model: str,
+        item1: Dict,
+        item2: Dict,
+        timeout_seconds: int = 120,
+        max_retries_per_timeout: int = 2,
+    ) -> Tuple[bool, float]:
+        """
+        Compares two items using an LLM model to determine if they match.
+
+        Args:
+            comparison_prompt (str): The prompt template for comparison.
+            model (str): The LLM model to use for comparison.
+            item1 (Dict): The first item to compare.
+            item2 (Dict): The second item to compare.
+            timeout_seconds (int): The timeout for the LLM call in seconds.
+            max_retries_per_timeout (int): The maximum number of retries per timeout.
+
+        Returns:
+            Tuple[bool, float]: A tuple containing a boolean indicating whether the items match and the cost of the comparison.
+        """
+
+        prompt_template = Template(comparison_prompt)
+        prompt = prompt_template.render(left=item1, right=item2)
+        response = self.runner.api.call_llm(
+            model,
+            "compare",
+            [{"role": "user", "content": prompt}],
+            {"is_match": "bool"},
+            timeout_seconds=timeout_seconds,
+            max_retries_per_timeout=max_retries_per_timeout,
+            bypass_cache=self.config.get("bypass_cache", False),
+            litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+        )
+        output = self.runner.api.parse_llm_response(
+            response.response, {"is_match": "bool"}
+        )[0]
+        return output["is_match"], response.total_cost
+
+    def syntax_check(self) -> None:
+        """
+        Checks the configuration of the EquijoinOperation for required keys and valid structure.
+
+        Raises:
+            ValueError: If required keys are missing or if the blocking_keys structure is invalid.
+            Specifically:
+            - Raises if 'comparison_prompt' is missing from the config.
+            - Raises if 'left' or 'right' are missing from the 'blocking_keys' structure (if present).
+            - Raises if 'left' or 'right' are missing from the 'limits' structure (if present).
+        """
+        if "comparison_prompt" not in self.config:
+            raise ValueError(
+                "Missing required key 'comparison_prompt' in EquijoinOperation configuration"
+            )
+
+        if "blocking_keys" in self.config:
+            if (
+                "left" not in self.config["blocking_keys"]
+                or "right" not in self.config["blocking_keys"]
+            ):
+                raise ValueError(
+                    "Both 'left' and 'right' must be specified in 'blocking_keys'"
+                )
+
+        if "limits" in self.config:
+            if (
+                "left" not in self.config["limits"]
+                or "right" not in self.config["limits"]
+            ):
+                raise ValueError(
+                    "Both 'left' and 'right' must be specified in 'limits'"
+                )
+
+        if "limit_comparisons" in self.config:
+            if not isinstance(self.config["limit_comparisons"], int):
+                raise ValueError("limit_comparisons must be an integer")
+
+    def execute(
+        self, left_data: List[Dict], right_data: List[Dict]
+    ) -> Tuple[List[Dict], float]:
+        """
+        Executes the equijoin operation on the provided datasets.
+
+        Args:
+            left_data (List[Dict]): The left dataset to join.
+            right_data (List[Dict]): The right dataset to join.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the joined results and the total cost of the operation.
+
+        Usage:
+        ```python
+        from docetl.operations import EquijoinOperation
+
+        config = {
+            "blocking_keys": {
+                "left": ["id"],
+                "right": ["user_id"]
+            },
+            "limits": {
+                "left": 1,
+                "right": 1
+            },
+            "comparison_prompt": "Compare {{left}} and {{right}} and determine if they match.",
+            "blocking_threshold": 0.8,
+            "blocking_conditions": ["left['id'] == right['user_id']"],
+            "limit_comparisons": 1000
+        }
+        equijoin_op = EquijoinOperation(config)
+        left_data = [{"id": 1, "name": "Alice"}, {"id": 2, "name": "Bob"}]
+        right_data = [{"user_id": 1, "age": 30}, {"user_id": 2, "age": 25}]
+        results, cost = equijoin_op.execute(left_data, right_data)
+        print(f"Joined results: {results}")
+        print(f"Total cost: {cost}")
+        ```
+
+        This method performs the following steps:
+        1. Initial blocking based on specified conditions (if any)
+        2. Embedding-based blocking (if threshold is provided)
+        3. LLM-based comparison for blocked pairs
+        4. Result aggregation and validation
+
+        The method also calculates and logs statistics such as comparisons saved by blocking and join selectivity.
+        """
+
+        blocking_keys = self.config.get("blocking_keys", {})
+        left_keys = blocking_keys.get(
+            "left", list(left_data[0].keys()) if left_data else []
+        )
+        right_keys = blocking_keys.get(
+            "right", list(right_data[0].keys()) if right_data else []
+        )
+        limits = self.config.get(
+            "limits", {"left": float("inf"), "right": float("inf")}
+        )
+        left_limit = limits["left"]
+        right_limit = limits["right"]
+        blocking_threshold = self.config.get("blocking_threshold")
+        blocking_conditions = self.config.get("blocking_conditions", [])
+        limit_comparisons = self.config.get("limit_comparisons")
+        total_cost = 0
+
+        # LLM-based comparison for blocked pairs
+        def get_hashable_key(item: Dict) -> str:
+            return json.dumps(item, sort_keys=True)
+
+        if len(left_data) == 0 or len(right_data) == 0:
+            return [], 0
+
+        if self.status:
+            self.status.stop()
+
+        # Initial blocking using multiprocessing
+        num_processes = min(cpu_count(), len(left_data))
+
+        self.console.log(
+            f"Starting to run code-based blocking rules for {len(left_data)} left and {len(right_data)} right rows ({len(left_data) * len(right_data)} total pairs) with {num_processes} processes..."
+        )
+
+        with Pool(
+            processes=num_processes,
+            initializer=init_worker,
+            initargs=(right_data, blocking_conditions),
+        ) as pool:
+            blocked_pairs_nested = pool.map(process_left_item, left_data)
+
+        # Flatten the nested list of blocked pairs
+        blocked_pairs = [pair for sublist in blocked_pairs_nested for pair in sublist]
+
+        # Check if we have exceeded the pairwise comparison limit
+        if limit_comparisons is not None and len(blocked_pairs) > limit_comparisons:
+            # Sample pairs randomly
+            sampled_pairs = random.sample(blocked_pairs, limit_comparisons)
+
+            # Calculate number of dropped pairs
+            dropped_pairs = len(blocked_pairs) - limit_comparisons
+
+            # Prompt the user for confirmation
+            if self.status:
+                self.status.stop()
+            if not Confirm.ask(
+                f"[yellow]Warning: {dropped_pairs} pairs will be dropped due to the comparison limit. "
+                f"Proceeding with {limit_comparisons} randomly sampled pairs. "
+                f"Do you want to continue?[/yellow]",
+                self.console,
+            ):
+                raise ValueError("Operation cancelled by user due to pair limit.")
+
+            if self.status:
+                self.status.start()
+
+            blocked_pairs = sampled_pairs
+
+        self.console.log(
+            f"Number of blocked pairs after initial blocking: {len(blocked_pairs)}"
+        )
+
+        if blocking_threshold is not None:
+            embedding_model = self.config.get("embedding_model", self.default_model)
+            model_input_context_length = model_cost.get(embedding_model, {}).get(
+                "max_input_tokens", 8192
+            )
+
+            def get_embeddings(
+                input_data: List[Dict[str, Any]], keys: List[str], name: str
+            ) -> Tuple[List[List[float]], float]:
+                texts = [
+                    " ".join(str(item[key]) for key in keys if key in item)[
+                        : model_input_context_length * 4
+                    ]
+                    for item in input_data
+                ]
+
+                embeddings = []
+                total_cost = 0
+                batch_size = 2000
+                for i in range(0, len(texts), batch_size):
+                    batch = texts[i : i + batch_size]
+                    self.console.log(
+                        f"On iteration {i} for creating embeddings for {name} data"
+                    )
+                    response = self.runner.api.gen_embedding(
+                        model=embedding_model,
+                        input=batch,
+                    )
+                    embeddings.extend([data["embedding"] for data in response["data"]])
+                    total_cost += completion_cost(response)
+                return embeddings, total_cost
+
+            left_embeddings, left_cost = get_embeddings(left_data, left_keys, "left")
+            right_embeddings, right_cost = get_embeddings(
+                right_data, right_keys, "right"
+            )
+            total_cost += left_cost + right_cost
+            self.console.log(
+                f"Created embeddings for datasets. Total embedding creation cost: {total_cost}"
+            )
+
+            # Compute all cosine similarities in one call
+            from sklearn.metrics.pairwise import cosine_similarity
+
+            similarities = cosine_similarity(left_embeddings, right_embeddings)
+
+            # Additional blocking based on embeddings
+            # Find indices where similarity is above threshold
+            above_threshold = np.argwhere(similarities >= blocking_threshold)
+            self.console.log(
+                f"There are {above_threshold.shape[0]} pairs above the threshold."
+            )
+            block_pair_set = set(
+                (get_hashable_key(left_item), get_hashable_key(right_item))
+                for left_item, right_item in blocked_pairs
+            )
+
+            # If limit_comparisons is set, take only the top pairs
+            if limit_comparisons is not None:
+                # First, get all pairs above threshold
+                above_threshold_pairs = [(int(i), int(j)) for i, j in above_threshold]
+
+                # Sort these pairs by their similarity scores
+                sorted_pairs = sorted(
+                    above_threshold_pairs,
+                    key=lambda pair: similarities[pair[0], pair[1]],
+                    reverse=True,
+                )
+
+                # Take the top 'limit_comparisons' pairs
+                top_pairs = sorted_pairs[:limit_comparisons]
+
+                # Create new blocked_pairs based on top similarities and existing blocked pairs
+                new_blocked_pairs = []
+                remaining_limit = limit_comparisons - len(blocked_pairs)
+
+                # First, include all existing blocked pairs
+                final_blocked_pairs = blocked_pairs.copy()
+
+                # Then, add new pairs from top similarities until we reach the limit
+                for i, j in top_pairs:
+                    if remaining_limit <= 0:
+                        break
+                    left_item, right_item = left_data[i], right_data[j]
+                    left_key = get_hashable_key(left_item)
+                    right_key = get_hashable_key(right_item)
+                    if (left_key, right_key) not in block_pair_set:
+                        new_blocked_pairs.append((left_item, right_item))
+                        block_pair_set.add((left_key, right_key))
+                        remaining_limit -= 1
+
+                final_blocked_pairs.extend(new_blocked_pairs)
+                blocked_pairs = final_blocked_pairs
+
+                self.console.log(
+                    f"Limited comparisons to top {limit_comparisons} pairs, including {len(blocked_pairs) - len(new_blocked_pairs)} from code-based blocking and {len(new_blocked_pairs)} based on cosine similarity. Lowest cosine similarity included: {similarities[top_pairs[-1]]:.4f}"
+                )
+            else:
+                # Add new pairs to blocked_pairs
+                for i, j in above_threshold:
+                    left_item, right_item = left_data[i], right_data[j]
+                    left_key = get_hashable_key(left_item)
+                    right_key = get_hashable_key(right_item)
+                    if (left_key, right_key) not in block_pair_set:
+                        blocked_pairs.append((left_item, right_item))
+                        block_pair_set.add((left_key, right_key))
+
+        # If there are no blocking conditions or embedding threshold, use all pairs
+        if not blocking_conditions and blocking_threshold is None:
+            blocked_pairs = [
+                (left_item, right_item)
+                for left_item in left_data
+                for right_item in right_data
+            ]
+
+        # If there's a limit on the number of comparisons, randomly sample pairs
+        if limit_comparisons is not None and len(blocked_pairs) > limit_comparisons:
+            self.console.log(
+                f"Randomly sampling {limit_comparisons} pairs out of {len(blocked_pairs)} blocked pairs."
+            )
+            blocked_pairs = random.sample(blocked_pairs, limit_comparisons)
+
+        self.console.log(
+            f"Total pairs to compare after blocking and sampling: {len(blocked_pairs)}"
+        )
+
+        # Calculate and print statistics
+        total_possible_comparisons = len(left_data) * len(right_data)
+        comparisons_made = len(blocked_pairs)
+        comparisons_saved = total_possible_comparisons - comparisons_made
+        self.console.log(
+            f"[green]Comparisons saved by blocking: {comparisons_saved} "
+            f"({(comparisons_saved / total_possible_comparisons) * 100:.2f}%)[/green]"
+        )
+
+        left_match_counts = defaultdict(int)
+        right_match_counts = defaultdict(int)
+        results = []
+        comparison_costs = 0
+
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            future_to_pair = {
+                executor.submit(
+                    self.compare_pair,
+                    self.config["comparison_prompt"],
+                    self.config.get("comparison_model", self.default_model),
+                    left,
+                    right,
+                    self.config.get("timeout", 120),
+                    self.config.get("max_retries_per_timeout", 2),
+                ): (left, right)
+                for left, right in blocked_pairs
+            }
+
+            for future in rich_as_completed(
+                future_to_pair,
+                total=len(future_to_pair),
+                desc="Comparing pairs",
+                console=self.console,
+            ):
+                pair = future_to_pair[future]
+                is_match, cost = future.result()
+                comparison_costs += cost
+
+                if is_match:
+                    joined_item = {}
+                    left_item, right_item = pair
+                    left_key_hash = get_hashable_key(left_item)
+                    right_key_hash = get_hashable_key(right_item)
+                    if (
+                        left_match_counts[left_key_hash] >= left_limit
+                        or right_match_counts[right_key_hash] >= right_limit
+                    ):
+                        continue
+
+                    for key, value in left_item.items():
+                        joined_item[f"{key}_left" if key in right_item else key] = value
+                    for key, value in right_item.items():
+                        joined_item[f"{key}_right" if key in left_item else key] = value
+                    if self.runner.api.validate_output(
+                        self.config, joined_item, self.console
+                    ):
+                        results.append(joined_item)
+                        left_match_counts[left_key_hash] += 1
+                        right_match_counts[right_key_hash] += 1
+
+                    # TODO: support retry in validation failure
+
+        total_cost += comparison_costs
+
+        # Calculate and print the join selectivity
+        join_selectivity = (
+            len(results) / (len(left_data) * len(right_data))
+            if len(left_data) * len(right_data) > 0
+            else 0
+        )
+        self.console.log(f"Equijoin selectivity: {join_selectivity:.4f}")
+
+        if self.status:
+            self.status.start()
+
+        return results, total_cost
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ compare_pair(comparison_prompt, model, item1, item2, timeout_seconds=120, max_retries_per_timeout=2) + +

+ + +
+ +

Compares two items using an LLM model to determine if they match.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ comparison_prompt + + str + +
+

The prompt template for comparison.

+
+
+ required +
+ model + + str + +
+

The LLM model to use for comparison.

+
+
+ required +
+ item1 + + Dict + +
+

The first item to compare.

+
+
+ required +
+ item2 + + Dict + +
+

The second item to compare.

+
+
+ required +
+ timeout_seconds + + int + +
+

The timeout for the LLM call in seconds.

+
+
+ 120 +
+ max_retries_per_timeout + + int + +
+

The maximum number of retries per timeout.

+
+
+ 2 +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[bool, float] + +
+

Tuple[bool, float]: A tuple containing a boolean indicating whether the items match and the cost of the comparison.

+
+
+ +
+ Source code in docetl/operations/equijoin.py +
 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
def compare_pair(
+    self,
+    comparison_prompt: str,
+    model: str,
+    item1: Dict,
+    item2: Dict,
+    timeout_seconds: int = 120,
+    max_retries_per_timeout: int = 2,
+) -> Tuple[bool, float]:
+    """
+    Compares two items using an LLM model to determine if they match.
+
+    Args:
+        comparison_prompt (str): The prompt template for comparison.
+        model (str): The LLM model to use for comparison.
+        item1 (Dict): The first item to compare.
+        item2 (Dict): The second item to compare.
+        timeout_seconds (int): The timeout for the LLM call in seconds.
+        max_retries_per_timeout (int): The maximum number of retries per timeout.
+
+    Returns:
+        Tuple[bool, float]: A tuple containing a boolean indicating whether the items match and the cost of the comparison.
+    """
+
+    prompt_template = Template(comparison_prompt)
+    prompt = prompt_template.render(left=item1, right=item2)
+    response = self.runner.api.call_llm(
+        model,
+        "compare",
+        [{"role": "user", "content": prompt}],
+        {"is_match": "bool"},
+        timeout_seconds=timeout_seconds,
+        max_retries_per_timeout=max_retries_per_timeout,
+        bypass_cache=self.config.get("bypass_cache", False),
+        litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+    )
+    output = self.runner.api.parse_llm_response(
+        response.response, {"is_match": "bool"}
+    )[0]
+    return output["is_match"], response.total_cost
+
+
+
+ +
+ +
+ + +

+ execute(left_data, right_data) + +

+ + +
+ +

Executes the equijoin operation on the provided datasets.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ left_data + + List[Dict] + +
+

The left dataset to join.

+
+
+ required +
+ right_data + + List[Dict] + +
+

The right dataset to join.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[List[Dict], float] + +
+

Tuple[List[Dict], float]: A tuple containing the joined results and the total cost of the operation.

+
+
+

Usage: +

from docetl.operations import EquijoinOperation
+
+config = {
+    "blocking_keys": {
+        "left": ["id"],
+        "right": ["user_id"]
+    },
+    "limits": {
+        "left": 1,
+        "right": 1
+    },
+    "comparison_prompt": "Compare {{left}} and {{right}} and determine if they match.",
+    "blocking_threshold": 0.8,
+    "blocking_conditions": ["left['id'] == right['user_id']"],
+    "limit_comparisons": 1000
+}
+equijoin_op = EquijoinOperation(config)
+left_data = [{"id": 1, "name": "Alice"}, {"id": 2, "name": "Bob"}]
+right_data = [{"user_id": 1, "age": 30}, {"user_id": 2, "age": 25}]
+results, cost = equijoin_op.execute(left_data, right_data)
+print(f"Joined results: {results}")
+print(f"Total cost: {cost}")
+

+

This method performs the following steps: +1. Initial blocking based on specified conditions (if any) +2. Embedding-based blocking (if threshold is provided) +3. LLM-based comparison for blocked pairs +4. Result aggregation and validation

+

The method also calculates and logs statistics such as comparisons saved by blocking and join selectivity.

+ +
+ Source code in docetl/operations/equijoin.py +
152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
+470
+471
+472
+473
def execute(
+    self, left_data: List[Dict], right_data: List[Dict]
+) -> Tuple[List[Dict], float]:
+    """
+    Executes the equijoin operation on the provided datasets.
+
+    Args:
+        left_data (List[Dict]): The left dataset to join.
+        right_data (List[Dict]): The right dataset to join.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the joined results and the total cost of the operation.
+
+    Usage:
+    ```python
+    from docetl.operations import EquijoinOperation
+
+    config = {
+        "blocking_keys": {
+            "left": ["id"],
+            "right": ["user_id"]
+        },
+        "limits": {
+            "left": 1,
+            "right": 1
+        },
+        "comparison_prompt": "Compare {{left}} and {{right}} and determine if they match.",
+        "blocking_threshold": 0.8,
+        "blocking_conditions": ["left['id'] == right['user_id']"],
+        "limit_comparisons": 1000
+    }
+    equijoin_op = EquijoinOperation(config)
+    left_data = [{"id": 1, "name": "Alice"}, {"id": 2, "name": "Bob"}]
+    right_data = [{"user_id": 1, "age": 30}, {"user_id": 2, "age": 25}]
+    results, cost = equijoin_op.execute(left_data, right_data)
+    print(f"Joined results: {results}")
+    print(f"Total cost: {cost}")
+    ```
+
+    This method performs the following steps:
+    1. Initial blocking based on specified conditions (if any)
+    2. Embedding-based blocking (if threshold is provided)
+    3. LLM-based comparison for blocked pairs
+    4. Result aggregation and validation
+
+    The method also calculates and logs statistics such as comparisons saved by blocking and join selectivity.
+    """
+
+    blocking_keys = self.config.get("blocking_keys", {})
+    left_keys = blocking_keys.get(
+        "left", list(left_data[0].keys()) if left_data else []
+    )
+    right_keys = blocking_keys.get(
+        "right", list(right_data[0].keys()) if right_data else []
+    )
+    limits = self.config.get(
+        "limits", {"left": float("inf"), "right": float("inf")}
+    )
+    left_limit = limits["left"]
+    right_limit = limits["right"]
+    blocking_threshold = self.config.get("blocking_threshold")
+    blocking_conditions = self.config.get("blocking_conditions", [])
+    limit_comparisons = self.config.get("limit_comparisons")
+    total_cost = 0
+
+    # LLM-based comparison for blocked pairs
+    def get_hashable_key(item: Dict) -> str:
+        return json.dumps(item, sort_keys=True)
+
+    if len(left_data) == 0 or len(right_data) == 0:
+        return [], 0
+
+    if self.status:
+        self.status.stop()
+
+    # Initial blocking using multiprocessing
+    num_processes = min(cpu_count(), len(left_data))
+
+    self.console.log(
+        f"Starting to run code-based blocking rules for {len(left_data)} left and {len(right_data)} right rows ({len(left_data) * len(right_data)} total pairs) with {num_processes} processes..."
+    )
+
+    with Pool(
+        processes=num_processes,
+        initializer=init_worker,
+        initargs=(right_data, blocking_conditions),
+    ) as pool:
+        blocked_pairs_nested = pool.map(process_left_item, left_data)
+
+    # Flatten the nested list of blocked pairs
+    blocked_pairs = [pair for sublist in blocked_pairs_nested for pair in sublist]
+
+    # Check if we have exceeded the pairwise comparison limit
+    if limit_comparisons is not None and len(blocked_pairs) > limit_comparisons:
+        # Sample pairs randomly
+        sampled_pairs = random.sample(blocked_pairs, limit_comparisons)
+
+        # Calculate number of dropped pairs
+        dropped_pairs = len(blocked_pairs) - limit_comparisons
+
+        # Prompt the user for confirmation
+        if self.status:
+            self.status.stop()
+        if not Confirm.ask(
+            f"[yellow]Warning: {dropped_pairs} pairs will be dropped due to the comparison limit. "
+            f"Proceeding with {limit_comparisons} randomly sampled pairs. "
+            f"Do you want to continue?[/yellow]",
+            self.console,
+        ):
+            raise ValueError("Operation cancelled by user due to pair limit.")
+
+        if self.status:
+            self.status.start()
+
+        blocked_pairs = sampled_pairs
+
+    self.console.log(
+        f"Number of blocked pairs after initial blocking: {len(blocked_pairs)}"
+    )
+
+    if blocking_threshold is not None:
+        embedding_model = self.config.get("embedding_model", self.default_model)
+        model_input_context_length = model_cost.get(embedding_model, {}).get(
+            "max_input_tokens", 8192
+        )
+
+        def get_embeddings(
+            input_data: List[Dict[str, Any]], keys: List[str], name: str
+        ) -> Tuple[List[List[float]], float]:
+            texts = [
+                " ".join(str(item[key]) for key in keys if key in item)[
+                    : model_input_context_length * 4
+                ]
+                for item in input_data
+            ]
+
+            embeddings = []
+            total_cost = 0
+            batch_size = 2000
+            for i in range(0, len(texts), batch_size):
+                batch = texts[i : i + batch_size]
+                self.console.log(
+                    f"On iteration {i} for creating embeddings for {name} data"
+                )
+                response = self.runner.api.gen_embedding(
+                    model=embedding_model,
+                    input=batch,
+                )
+                embeddings.extend([data["embedding"] for data in response["data"]])
+                total_cost += completion_cost(response)
+            return embeddings, total_cost
+
+        left_embeddings, left_cost = get_embeddings(left_data, left_keys, "left")
+        right_embeddings, right_cost = get_embeddings(
+            right_data, right_keys, "right"
+        )
+        total_cost += left_cost + right_cost
+        self.console.log(
+            f"Created embeddings for datasets. Total embedding creation cost: {total_cost}"
+        )
+
+        # Compute all cosine similarities in one call
+        from sklearn.metrics.pairwise import cosine_similarity
+
+        similarities = cosine_similarity(left_embeddings, right_embeddings)
+
+        # Additional blocking based on embeddings
+        # Find indices where similarity is above threshold
+        above_threshold = np.argwhere(similarities >= blocking_threshold)
+        self.console.log(
+            f"There are {above_threshold.shape[0]} pairs above the threshold."
+        )
+        block_pair_set = set(
+            (get_hashable_key(left_item), get_hashable_key(right_item))
+            for left_item, right_item in blocked_pairs
+        )
+
+        # If limit_comparisons is set, take only the top pairs
+        if limit_comparisons is not None:
+            # First, get all pairs above threshold
+            above_threshold_pairs = [(int(i), int(j)) for i, j in above_threshold]
+
+            # Sort these pairs by their similarity scores
+            sorted_pairs = sorted(
+                above_threshold_pairs,
+                key=lambda pair: similarities[pair[0], pair[1]],
+                reverse=True,
+            )
+
+            # Take the top 'limit_comparisons' pairs
+            top_pairs = sorted_pairs[:limit_comparisons]
+
+            # Create new blocked_pairs based on top similarities and existing blocked pairs
+            new_blocked_pairs = []
+            remaining_limit = limit_comparisons - len(blocked_pairs)
+
+            # First, include all existing blocked pairs
+            final_blocked_pairs = blocked_pairs.copy()
+
+            # Then, add new pairs from top similarities until we reach the limit
+            for i, j in top_pairs:
+                if remaining_limit <= 0:
+                    break
+                left_item, right_item = left_data[i], right_data[j]
+                left_key = get_hashable_key(left_item)
+                right_key = get_hashable_key(right_item)
+                if (left_key, right_key) not in block_pair_set:
+                    new_blocked_pairs.append((left_item, right_item))
+                    block_pair_set.add((left_key, right_key))
+                    remaining_limit -= 1
+
+            final_blocked_pairs.extend(new_blocked_pairs)
+            blocked_pairs = final_blocked_pairs
+
+            self.console.log(
+                f"Limited comparisons to top {limit_comparisons} pairs, including {len(blocked_pairs) - len(new_blocked_pairs)} from code-based blocking and {len(new_blocked_pairs)} based on cosine similarity. Lowest cosine similarity included: {similarities[top_pairs[-1]]:.4f}"
+            )
+        else:
+            # Add new pairs to blocked_pairs
+            for i, j in above_threshold:
+                left_item, right_item = left_data[i], right_data[j]
+                left_key = get_hashable_key(left_item)
+                right_key = get_hashable_key(right_item)
+                if (left_key, right_key) not in block_pair_set:
+                    blocked_pairs.append((left_item, right_item))
+                    block_pair_set.add((left_key, right_key))
+
+    # If there are no blocking conditions or embedding threshold, use all pairs
+    if not blocking_conditions and blocking_threshold is None:
+        blocked_pairs = [
+            (left_item, right_item)
+            for left_item in left_data
+            for right_item in right_data
+        ]
+
+    # If there's a limit on the number of comparisons, randomly sample pairs
+    if limit_comparisons is not None and len(blocked_pairs) > limit_comparisons:
+        self.console.log(
+            f"Randomly sampling {limit_comparisons} pairs out of {len(blocked_pairs)} blocked pairs."
+        )
+        blocked_pairs = random.sample(blocked_pairs, limit_comparisons)
+
+    self.console.log(
+        f"Total pairs to compare after blocking and sampling: {len(blocked_pairs)}"
+    )
+
+    # Calculate and print statistics
+    total_possible_comparisons = len(left_data) * len(right_data)
+    comparisons_made = len(blocked_pairs)
+    comparisons_saved = total_possible_comparisons - comparisons_made
+    self.console.log(
+        f"[green]Comparisons saved by blocking: {comparisons_saved} "
+        f"({(comparisons_saved / total_possible_comparisons) * 100:.2f}%)[/green]"
+    )
+
+    left_match_counts = defaultdict(int)
+    right_match_counts = defaultdict(int)
+    results = []
+    comparison_costs = 0
+
+    with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+        future_to_pair = {
+            executor.submit(
+                self.compare_pair,
+                self.config["comparison_prompt"],
+                self.config.get("comparison_model", self.default_model),
+                left,
+                right,
+                self.config.get("timeout", 120),
+                self.config.get("max_retries_per_timeout", 2),
+            ): (left, right)
+            for left, right in blocked_pairs
+        }
+
+        for future in rich_as_completed(
+            future_to_pair,
+            total=len(future_to_pair),
+            desc="Comparing pairs",
+            console=self.console,
+        ):
+            pair = future_to_pair[future]
+            is_match, cost = future.result()
+            comparison_costs += cost
+
+            if is_match:
+                joined_item = {}
+                left_item, right_item = pair
+                left_key_hash = get_hashable_key(left_item)
+                right_key_hash = get_hashable_key(right_item)
+                if (
+                    left_match_counts[left_key_hash] >= left_limit
+                    or right_match_counts[right_key_hash] >= right_limit
+                ):
+                    continue
+
+                for key, value in left_item.items():
+                    joined_item[f"{key}_left" if key in right_item else key] = value
+                for key, value in right_item.items():
+                    joined_item[f"{key}_right" if key in left_item else key] = value
+                if self.runner.api.validate_output(
+                    self.config, joined_item, self.console
+                ):
+                    results.append(joined_item)
+                    left_match_counts[left_key_hash] += 1
+                    right_match_counts[right_key_hash] += 1
+
+                # TODO: support retry in validation failure
+
+    total_cost += comparison_costs
+
+    # Calculate and print the join selectivity
+    join_selectivity = (
+        len(results) / (len(left_data) * len(right_data))
+        if len(left_data) * len(right_data) > 0
+        else 0
+    )
+    self.console.log(f"Equijoin selectivity: {join_selectivity:.4f}")
+
+    if self.status:
+        self.status.start()
+
+    return results, total_cost
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Checks the configuration of the EquijoinOperation for required keys and valid structure.

+ + +

Raises:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If required keys are missing or if the blocking_keys structure is invalid.

+
+
+ Specifically + +
+ +
+
+ +
+ Source code in docetl/operations/equijoin.py +
114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
def syntax_check(self) -> None:
+    """
+    Checks the configuration of the EquijoinOperation for required keys and valid structure.
+
+    Raises:
+        ValueError: If required keys are missing or if the blocking_keys structure is invalid.
+        Specifically:
+        - Raises if 'comparison_prompt' is missing from the config.
+        - Raises if 'left' or 'right' are missing from the 'blocking_keys' structure (if present).
+        - Raises if 'left' or 'right' are missing from the 'limits' structure (if present).
+    """
+    if "comparison_prompt" not in self.config:
+        raise ValueError(
+            "Missing required key 'comparison_prompt' in EquijoinOperation configuration"
+        )
+
+    if "blocking_keys" in self.config:
+        if (
+            "left" not in self.config["blocking_keys"]
+            or "right" not in self.config["blocking_keys"]
+        ):
+            raise ValueError(
+                "Both 'left' and 'right' must be specified in 'blocking_keys'"
+            )
+
+    if "limits" in self.config:
+        if (
+            "left" not in self.config["limits"]
+            or "right" not in self.config["limits"]
+        ):
+            raise ValueError(
+                "Both 'left' and 'right' must be specified in 'limits'"
+            )
+
+    if "limit_comparisons" in self.config:
+        if not isinstance(self.config["limit_comparisons"], int):
+            raise ValueError("limit_comparisons must be an integer")
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.operations.cluster.ClusterOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + + + + + + +
+ Source code in docetl/operations/cluster.py +
 10
+ 11
+ 12
+ 13
+ 14
+ 15
+ 16
+ 17
+ 18
+ 19
+ 20
+ 21
+ 22
+ 23
+ 24
+ 25
+ 26
+ 27
+ 28
+ 29
+ 30
+ 31
+ 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
class ClusterOperation(BaseOperation):
+    def __init__(
+        self,
+        *args,
+        **kwargs,
+    ):
+        super().__init__(*args, **kwargs)
+        self.max_batch_size: int = self.config.get(
+            "max_batch_size", kwargs.get("max_batch_size", float("inf"))
+        )
+
+    def syntax_check(self) -> None:
+        """
+        Checks the configuration of the ClusterOperation for required keys and valid structure.
+
+        Raises:
+            ValueError: If required keys are missing or invalid in the configuration.
+            TypeError: If configuration values have incorrect types.
+        """
+        required_keys = ["embedding_keys", "summary_schema", "summary_prompt"]
+        for key in required_keys:
+            if key not in self.config:
+                raise ValueError(
+                    f"Missing required key '{key}' in ClusterOperation configuration"
+                )
+
+        if not isinstance(self.config["embedding_keys"], list):
+            raise TypeError("'embedding_keys' must be a list of strings")
+
+        if "output_key" in self.config:
+            if not isinstance(self.config["output_key"], str):
+                raise TypeError("'output_key' must be a string")
+
+        if not isinstance(self.config["summary_schema"], dict):
+            raise TypeError("'summary_schema' must be a dictionary")
+
+        if not isinstance(self.config["summary_prompt"], str):
+            raise TypeError("'prompt' must be a string")
+
+        # Check if the prompt is a valid Jinja2 template
+        try:
+            Template(self.config["summary_prompt"])
+        except Exception as e:
+            raise ValueError(f"Invalid Jinja2 template in 'prompt': {str(e)}")
+
+        # Check optional parameters
+        if "max_batch_size" in self.config:
+            if not isinstance(self.config["max_batch_size"], int):
+                raise TypeError("'max_batch_size' must be an integer")
+
+        if "embedding_model" in self.config:
+            if not isinstance(self.config["embedding_model"], str):
+                raise TypeError("'embedding_model' must be a string")
+
+        if "model" in self.config:
+            if not isinstance(self.config["model"], str):
+                raise TypeError("'model' must be a string")
+
+        if "validate" in self.config:
+            if not isinstance(self.config["validate"], list):
+                raise TypeError("'validate' must be a list of strings")
+            for rule in self.config["validate"]:
+                if not isinstance(rule, str):
+                    raise TypeError("Each validation rule must be a string")
+
+    def execute(
+        self, input_data: List[Dict], is_build: bool = False
+    ) -> Tuple[List[Dict], float]:
+        """
+        Executes the cluster operation on the input data. Modifies the
+        input data and returns it in place.
+
+        Args:
+            input_data (List[Dict]): A list of dictionaries to process.
+            is_build (bool): Whether the operation is being executed
+              in the build phase. Defaults to False.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the clustered
+              list of dictionaries and the total cost of the operation.
+        """
+        if not input_data:
+            return input_data, 0
+
+        if len(input_data) == 1:
+            input_data[0][self.config.get("output_key", "clusters")] = ()
+            return input_data, 0
+
+        embeddings, cost = get_embeddings_for_clustering(
+            input_data, self.config, self.runner.api
+        )
+
+        tree = self.agglomerative_cluster_of_embeddings(input_data, embeddings)
+
+        if "collapse" in self.config:
+            tree = self.collapse_tree(tree, collapse = self.config["collapse"])
+
+        self.prompt_template = Template(self.config["summary_prompt"])
+        cost += self.annotate_clustering_tree(tree)
+        self.annotate_leaves(tree)
+
+        return input_data, cost
+
+    def agglomerative_cluster_of_embeddings(self, input_data, embeddings):
+        import sklearn.cluster
+
+        cl = sklearn.cluster.AgglomerativeClustering(
+            compute_full_tree=True, compute_distances=True
+        )
+        cl.fit(embeddings)
+
+        nsamples = len(embeddings)
+
+        def build_tree(i):
+            if i < nsamples:
+                res = input_data[i]
+                #                res["embedding"] = list(embeddings[i])
+                return res
+            return {
+                 "children": [
+                    build_tree(cl.children_[i - nsamples, 0]),
+                    build_tree(cl.children_[i - nsamples, 1]),
+                ],
+                "distance": cl.distances_[i - nsamples],
+            }
+
+        return build_tree(nsamples + len(cl.children_) - 1)
+
+    def get_tree_distances(self, t):
+        res = set()
+        if "distance" in t:
+            res.update(set([t["distance"] - child["distance"] for child in t["children"] if "distance" in child]))
+        if "children" in t:
+            for child in t["children"]:
+                res.update(self.get_tree_distances(child))
+        return res
+
+    def _collapse_tree(self, t, parent_dist = None, collapse = None):
+        if "children" in t:
+            if (    "distance" in t
+                and parent_dist is not None
+                and collapse is not None
+                and parent_dist - t["distance"] < collapse):
+                return [grandchild
+                        for child in t["children"]
+                        for grandchild in self._collapse_tree(child, parent_dist=parent_dist, collapse=collapse)]
+            else:
+                res = dict(t)
+                res["children"] = [grandchild
+                                   for idx, child in enumerate(t["children"])
+                                   for grandchild in self._collapse_tree(child, parent_dist=t["distance"], collapse=collapse)]
+                return [res]
+        else:
+            return [t]
+
+    def collapse_tree(self, tree, collapse = None):
+        if collapse is not None:
+            tree_distances = np.array(sorted(self.get_tree_distances(tree)))
+            collapse = tree_distances[int(len(tree_distances) * collapse)]
+        return self._collapse_tree(tree, collapse=collapse)[0]
+
+
+    def annotate_clustering_tree(self, t):
+        if "children" in t:
+            with ThreadPoolExecutor(max_workers=self.max_batch_size) as executor:
+                futures = [
+                    executor.submit(self.annotate_clustering_tree, child)
+                    for child in t["children"]
+                ]
+
+                total_cost = 0
+                pbar = RichLoopBar(
+                    range(len(futures)),
+                    desc=f"Processing {self.config['name']} (map) on all documents",
+                    console=self.console,
+                )
+                for i in pbar:
+                    total_cost += futures[i].result()
+                    pbar.update(i)
+
+            prompt = self.prompt_template.render(
+                inputs=t["children"]
+            )
+
+            def validation_fn(response: Dict[str, Any]):
+                output = self.runner.api.parse_llm_response(
+                    response,
+                    schema=self.config["summary_schema"],
+                    manually_fix_errors=self.manually_fix_errors,
+                )[0]
+                if self.runner.api.validate_output(self.config, output, self.console):
+                    return output, True
+                return output, False
+
+            response = self.runner.api.call_llm(
+                model=self.config.get("model", self.default_model),
+                op_type="cluster",
+                messages=[{"role": "user", "content": prompt}],
+                output_schema=self.config["summary_schema"],
+                timeout_seconds=self.config.get("timeout", 120),
+                bypass_cache=self.config.get("bypass_cache", False),
+                max_retries_per_timeout=self.config.get("max_retries_per_timeout", 2),
+                validation_config=(
+                    {
+                        "num_retries": self.num_retries_on_validate_failure,
+                        "val_rule": self.config.get("validate", []),
+                        "validation_fn": validation_fn,
+                    }
+                    if self.config.get("validate", None)
+                    else None
+                ),
+                verbose=self.config.get("verbose", False),
+                litellm_completion_kwargs=self.config.get("litellm_completion_kwargs", {}),
+            )
+            total_cost += response.total_cost
+            if response.validated:
+                output = self.runner.api.parse_llm_response(
+                    response.response,
+                    schema=self.config["summary_schema"],
+                    manually_fix_errors=self.manually_fix_errors,
+                )[0]
+                t.update(output)
+
+            return total_cost
+        return 0
+
+    def annotate_leaves(self, tree, path=()):
+        if "children" in tree:
+            item = dict(tree)
+            item.pop("children")
+            for child in tree["children"]:
+                self.annotate_leaves(child, path=(item,) + path)
+        else:
+            tree[self.config.get("output_key", "clusters")] = path
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ execute(input_data, is_build=False) + +

+ + +
+ +

Executes the cluster operation on the input data. Modifies the +input data and returns it in place.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ input_data + + List[Dict] + +
+

A list of dictionaries to process.

+
+
+ required +
+ is_build + + bool + +
+

Whether the operation is being executed +in the build phase. Defaults to False.

+
+
+ False +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[List[Dict], float] + +
+

Tuple[List[Dict], float]: A tuple containing the clustered +list of dictionaries and the total cost of the operation.

+
+
+ +
+ Source code in docetl/operations/cluster.py +
 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
def execute(
+    self, input_data: List[Dict], is_build: bool = False
+) -> Tuple[List[Dict], float]:
+    """
+    Executes the cluster operation on the input data. Modifies the
+    input data and returns it in place.
+
+    Args:
+        input_data (List[Dict]): A list of dictionaries to process.
+        is_build (bool): Whether the operation is being executed
+          in the build phase. Defaults to False.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the clustered
+          list of dictionaries and the total cost of the operation.
+    """
+    if not input_data:
+        return input_data, 0
+
+    if len(input_data) == 1:
+        input_data[0][self.config.get("output_key", "clusters")] = ()
+        return input_data, 0
+
+    embeddings, cost = get_embeddings_for_clustering(
+        input_data, self.config, self.runner.api
+    )
+
+    tree = self.agglomerative_cluster_of_embeddings(input_data, embeddings)
+
+    if "collapse" in self.config:
+        tree = self.collapse_tree(tree, collapse = self.config["collapse"])
+
+    self.prompt_template = Template(self.config["summary_prompt"])
+    cost += self.annotate_clustering_tree(tree)
+    self.annotate_leaves(tree)
+
+    return input_data, cost
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Checks the configuration of the ClusterOperation for required keys and valid structure.

+ + +

Raises:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If required keys are missing or invalid in the configuration.

+
+
+ TypeError + +
+

If configuration values have incorrect types.

+
+
+ +
+ Source code in docetl/operations/cluster.py +
21
+22
+23
+24
+25
+26
+27
+28
+29
+30
+31
+32
+33
+34
+35
+36
+37
+38
+39
+40
+41
+42
+43
+44
+45
+46
+47
+48
+49
+50
+51
+52
+53
+54
+55
+56
+57
+58
+59
+60
+61
+62
+63
+64
+65
+66
+67
+68
+69
+70
+71
+72
+73
def syntax_check(self) -> None:
+    """
+    Checks the configuration of the ClusterOperation for required keys and valid structure.
+
+    Raises:
+        ValueError: If required keys are missing or invalid in the configuration.
+        TypeError: If configuration values have incorrect types.
+    """
+    required_keys = ["embedding_keys", "summary_schema", "summary_prompt"]
+    for key in required_keys:
+        if key not in self.config:
+            raise ValueError(
+                f"Missing required key '{key}' in ClusterOperation configuration"
+            )
+
+    if not isinstance(self.config["embedding_keys"], list):
+        raise TypeError("'embedding_keys' must be a list of strings")
+
+    if "output_key" in self.config:
+        if not isinstance(self.config["output_key"], str):
+            raise TypeError("'output_key' must be a string")
+
+    if not isinstance(self.config["summary_schema"], dict):
+        raise TypeError("'summary_schema' must be a dictionary")
+
+    if not isinstance(self.config["summary_prompt"], str):
+        raise TypeError("'prompt' must be a string")
+
+    # Check if the prompt is a valid Jinja2 template
+    try:
+        Template(self.config["summary_prompt"])
+    except Exception as e:
+        raise ValueError(f"Invalid Jinja2 template in 'prompt': {str(e)}")
+
+    # Check optional parameters
+    if "max_batch_size" in self.config:
+        if not isinstance(self.config["max_batch_size"], int):
+            raise TypeError("'max_batch_size' must be an integer")
+
+    if "embedding_model" in self.config:
+        if not isinstance(self.config["embedding_model"], str):
+            raise TypeError("'embedding_model' must be a string")
+
+    if "model" in self.config:
+        if not isinstance(self.config["model"], str):
+            raise TypeError("'model' must be a string")
+
+    if "validate" in self.config:
+        if not isinstance(self.config["validate"], list):
+            raise TypeError("'validate' must be a list of strings")
+        for rule in self.config["validate"]:
+            if not isinstance(rule, str):
+                raise TypeError("Each validation rule must be a string")
+
+
+
+ +
+ + + +
+ +
+ +

Auxiliary Operators

+ + +
+ + + +

+ docetl.operations.split.SplitOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + +

A class that implements a split operation on input data, dividing it into manageable chunks.

+

This class extends BaseOperation to: +1. Split input data into chunks of specified size based on the 'split_key' and 'token_count' configuration. +2. Assign unique identifiers to each original document and number chunks sequentially. +3. Return results containing: + - {split_key}_chunk: The content of the split chunk. + - {name}_id: A unique identifier for each original document. + - {name}_chunk_num: The sequential number of the chunk within its original document.

+ + + + + + +
+ Source code in docetl/operations/split.py +
  9
+ 10
+ 11
+ 12
+ 13
+ 14
+ 15
+ 16
+ 17
+ 18
+ 19
+ 20
+ 21
+ 22
+ 23
+ 24
+ 25
+ 26
+ 27
+ 28
+ 29
+ 30
+ 31
+ 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
class SplitOperation(BaseOperation):
+    """
+    A class that implements a split operation on input data, dividing it into manageable chunks.
+
+    This class extends BaseOperation to:
+    1. Split input data into chunks of specified size based on the 'split_key' and 'token_count' configuration.
+    2. Assign unique identifiers to each original document and number chunks sequentially.
+    3. Return results containing:
+       - {split_key}_chunk: The content of the split chunk.
+       - {name}_id: A unique identifier for each original document.
+       - {name}_chunk_num: The sequential number of the chunk within its original document.
+    """
+    class schema(BaseOperation.schema):
+        type: str = "split"
+        split_key: str
+        method: str
+        method_kwargs: Dict[str, Any]
+        model: Optional[str] = None
+
+    def __init__(self, *args, **kwargs):
+        super().__init__(*args, **kwargs)
+        self.name = self.config["name"]
+
+    def syntax_check(self) -> None:
+        required_keys = ["split_key", "method", "method_kwargs"]
+        for key in required_keys:
+            if key not in self.config:
+                raise ValueError(
+                    f"Missing required key '{key}' in SplitOperation configuration"
+                )
+
+        if not isinstance(self.config["split_key"], str):
+            raise TypeError("'split_key' must be a string")
+
+        if self.config["method"] not in ["token_count", "delimiter"]:
+            raise ValueError(f"Invalid method '{self.config['method']}'")
+
+        if self.config["method"] == "token_count":
+            if (
+                not isinstance(self.config["method_kwargs"]["num_tokens"], int)
+                or self.config["method_kwargs"]["num_tokens"] <= 0
+            ):
+                raise ValueError("'num_tokens' must be a positive integer")
+        elif self.config["method"] == "delimiter":
+            if not isinstance(self.config["method_kwargs"]["delimiter"], str):
+                raise ValueError("'delimiter' must be a string")
+
+    def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+        split_key = self.config["split_key"]
+        method = self.config["method"]
+        method_kwargs = self.config["method_kwargs"]
+        try:
+            encoder = tiktoken.encoding_for_model(
+                self.config["method_kwargs"]
+                .get("model", self.default_model)
+                .split("/")[-1]
+            )
+        except Exception:
+            encoder = tiktoken.encoding_for_model("gpt-4o")
+
+        results = []
+        cost = 0.0
+
+        for item in input_data:
+            if split_key not in item:
+                raise KeyError(f"Split key '{split_key}' not found in item")
+
+            content = item[split_key]
+            doc_id = str(uuid.uuid4())
+
+            if method == "token_count":
+                token_count = method_kwargs["num_tokens"]
+                tokens = encoder.encode(content)
+
+                for chunk_num, i in enumerate(
+                    range(0, len(tokens), token_count), start=1
+                ):
+                    chunk_tokens = tokens[i : i + token_count]
+                    chunk = encoder.decode(chunk_tokens)
+
+                    result = item.copy()
+                    result.update(
+                        {
+                            f"{split_key}_chunk": chunk,
+                            f"{self.name}_id": doc_id,
+                            f"{self.name}_chunk_num": chunk_num,
+                        }
+                    )
+                    results.append(result)
+
+            elif method == "delimiter":
+                delimiter = method_kwargs["delimiter"]
+                num_splits_to_group = method_kwargs.get("num_splits_to_group", 1)
+                chunks = content.split(delimiter)
+
+                # Get rid of empty chunks
+                chunks = [chunk for chunk in chunks if chunk.strip()]
+
+                for chunk_num, i in enumerate(
+                    range(0, len(chunks), num_splits_to_group), start=1
+                ):
+                    grouped_chunks = chunks[i : i + num_splits_to_group]
+                    joined_chunk = delimiter.join(grouped_chunks).strip()
+
+                    result = item.copy()
+                    result.update(
+                        {
+                            f"{split_key}_chunk": joined_chunk,
+                            f"{self.name}_id": doc_id,
+                            f"{self.name}_chunk_num": chunk_num,
+                        }
+                    )
+                    results.append(result)
+
+        return results, cost
+
+
+ + + +
+ + + + + + + + + + + +
+ +
+ +
+ +
+ + + +

+ docetl.operations.gather.GatherOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + +

A class that implements a gather operation on input data, adding contextual information from surrounding chunks.

+

This class extends BaseOperation to: +1. Group chunks by their document ID. +2. Order chunks within each group. +3. Add peripheral context to each chunk based on the configuration. +4. Include headers for each chunk and its upward hierarchy. +5. Return results containing the rendered chunks with added context, including information about skipped characters and headers.

+ + + + + + +
+ Source code in docetl/operations/gather.py +
  6
+  7
+  8
+  9
+ 10
+ 11
+ 12
+ 13
+ 14
+ 15
+ 16
+ 17
+ 18
+ 19
+ 20
+ 21
+ 22
+ 23
+ 24
+ 25
+ 26
+ 27
+ 28
+ 29
+ 30
+ 31
+ 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
class GatherOperation(BaseOperation):
+    """
+    A class that implements a gather operation on input data, adding contextual information from surrounding chunks.
+
+    This class extends BaseOperation to:
+    1. Group chunks by their document ID.
+    2. Order chunks within each group.
+    3. Add peripheral context to each chunk based on the configuration.
+    4. Include headers for each chunk and its upward hierarchy.
+    5. Return results containing the rendered chunks with added context, including information about skipped characters and headers.
+    """
+
+    class schema(BaseOperation.schema):
+        type: str = "gather"
+        content_key: str
+        doc_id_key: str
+        order_key: str
+        peripheral_chunks: Dict[str, Any]
+        doc_header_key: Optional[str] = None
+
+    def __init__(self, *args: Any, **kwargs: Any) -> None:
+        """
+        Initialize the GatherOperation.
+
+        Args:
+            *args: Variable length argument list.
+            **kwargs: Arbitrary keyword arguments.
+        """
+        super().__init__(*args, **kwargs)
+
+    def syntax_check(self) -> None:
+        """
+        Perform a syntax check on the operation configuration.
+
+        Raises:
+            ValueError: If required keys are missing or if there are configuration errors.
+            TypeError: If main_chunk_start or main_chunk_end are not strings.
+        """
+        required_keys = ["content_key", "doc_id_key", "order_key"]
+        for key in required_keys:
+            if key not in self.config:
+                raise ValueError(
+                    f"Missing required key '{key}' in GatherOperation configuration"
+                )
+
+        peripheral_config = self.config.get("peripheral_chunks", {})
+        for direction in ["previous", "next"]:
+            if direction not in peripheral_config:
+                continue
+            for section in ["head", "middle", "tail"]:
+                if section in peripheral_config[direction]:
+                    section_config = peripheral_config[direction][section]
+                    if section != "middle" and "count" not in section_config:
+                        raise ValueError(
+                            f"Missing 'count' in {direction}.{section} configuration"
+                        )
+
+        if "main_chunk_start" in self.config and not isinstance(
+            self.config["main_chunk_start"], str
+        ):
+            raise TypeError("'main_chunk_start' must be a string")
+        if "main_chunk_end" in self.config and not isinstance(
+            self.config["main_chunk_end"], str
+        ):
+            raise TypeError("'main_chunk_end' must be a string")
+
+    def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+        """
+        Execute the gather operation on the input data.
+
+        Args:
+            input_data (List[Dict]): The input data to process.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the processed results and the cost of the operation.
+        """
+        content_key = self.config["content_key"]
+        doc_id_key = self.config["doc_id_key"]
+        order_key = self.config["order_key"]
+        peripheral_config = self.config.get("peripheral_chunks", {})
+        main_chunk_start = self.config.get(
+            "main_chunk_start", "--- Begin Main Chunk ---"
+        )
+        main_chunk_end = self.config.get("main_chunk_end", "--- End Main Chunk ---")
+        doc_header_key = self.config.get("doc_header_key", None)
+        results = []
+        cost = 0.0
+
+        # Group chunks by document ID
+        grouped_chunks = {}
+        for item in input_data:
+            doc_id = item[doc_id_key]
+            if doc_id not in grouped_chunks:
+                grouped_chunks[doc_id] = []
+            grouped_chunks[doc_id].append(item)
+
+        # Process each group of chunks
+        for chunks in grouped_chunks.values():
+            # Sort chunks by their order within the document
+            chunks.sort(key=lambda x: x[order_key])
+
+            # Process each chunk with its peripheral context and headers
+            for i, chunk in enumerate(chunks):
+                rendered_chunk = self.render_chunk_with_context(
+                    chunks,
+                    i,
+                    peripheral_config,
+                    content_key,
+                    order_key,
+                    main_chunk_start,
+                    main_chunk_end,
+                    doc_header_key,
+                )
+
+                result = chunk.copy()
+                result[f"{content_key}_rendered"] = rendered_chunk
+                results.append(result)
+
+        return results, cost
+
+    def render_chunk_with_context(
+        self,
+        chunks: List[Dict],
+        current_index: int,
+        peripheral_config: Dict,
+        content_key: str,
+        order_key: str,
+        main_chunk_start: str,
+        main_chunk_end: str,
+        doc_header_key: str,
+    ) -> str:
+        """
+        Render a chunk with its peripheral context and headers.
+
+        Args:
+            chunks (List[Dict]): List of all chunks in the document.
+            current_index (int): Index of the current chunk being processed.
+            peripheral_config (Dict): Configuration for peripheral chunks.
+            content_key (str): Key for the content in each chunk.
+            order_key (str): Key for the order of each chunk.
+            main_chunk_start (str): String to mark the start of the main chunk.
+            main_chunk_end (str): String to mark the end of the main chunk.
+            doc_header_key (str): The key for the headers in the current chunk.
+
+        Returns:
+            str: Renderted chunk with context and headers.
+        """
+        combined_parts = ["--- Previous Context ---"]
+
+        combined_parts.extend(
+            self.process_peripheral_chunks(
+                chunks[:current_index],
+                peripheral_config.get("previous", {}),
+                content_key,
+                order_key,
+            )
+        )
+        combined_parts.append("--- End Previous Context ---\n")
+
+        # Process main chunk
+        main_chunk = chunks[current_index]
+        if headers := self.render_hierarchy_headers(
+            main_chunk, chunks[: current_index + 1], doc_header_key
+        ):
+            combined_parts.append(headers)
+        combined_parts.extend(
+            (
+                f"{main_chunk_start}",
+                f"{main_chunk[content_key]}",
+                f"{main_chunk_end}",
+                "\n--- Next Context ---",
+            )
+        )
+        combined_parts.extend(
+            self.process_peripheral_chunks(
+                chunks[current_index + 1 :],
+                peripheral_config.get("next", {}),
+                content_key,
+                order_key,
+            )
+        )
+        combined_parts.append("--- End Next Context ---")
+
+        return "\n".join(combined_parts)
+
+    def process_peripheral_chunks(
+        self,
+        chunks: List[Dict],
+        config: Dict,
+        content_key: str,
+        order_key: str,
+        reverse: bool = False,
+    ) -> List[str]:
+        """
+        Process peripheral chunks according to the configuration.
+
+        Args:
+            chunks (List[Dict]): List of chunks to process.
+            config (Dict): Configuration for processing peripheral chunks.
+            content_key (str): Key for the content in each chunk.
+            order_key (str): Key for the order of each chunk.
+            reverse (bool, optional): Whether to process chunks in reverse order. Defaults to False.
+
+        Returns:
+            List[str]: List of processed chunk strings.
+        """
+        if reverse:
+            chunks = list(reversed(chunks))
+
+        processed_parts = []
+        included_chunks = []
+        total_chunks = len(chunks)
+
+        head_config = config.get("head", {})
+        tail_config = config.get("tail", {})
+
+        head_count = int(head_config.get("count", 0))
+        tail_count = int(tail_config.get("count", 0))
+        in_skip = False
+        skip_char_count = 0
+
+        for i, chunk in enumerate(chunks):
+            if i < head_count:
+                section = "head"
+            elif i >= total_chunks - tail_count:
+                section = "tail"
+            elif "middle" in config:
+                section = "middle"
+            else:
+                # Show number of characters skipped
+                skipped_chars = len(chunk[content_key])
+                if not in_skip:
+                    skip_char_count = skipped_chars
+                    in_skip = True
+                else:
+                    skip_char_count += skipped_chars
+
+                continue
+
+            if in_skip:
+                processed_parts.append(
+                    f"[... {skip_char_count} characters skipped ...]"
+                )
+                in_skip = False
+                skip_char_count = 0
+
+            section_config = config.get(section, {})
+            section_content_key = section_config.get("content_key", content_key)
+
+            is_summary = section_content_key != content_key
+            summary_suffix = " (Summary)" if is_summary else ""
+
+            chunk_prefix = f"[Chunk {chunk[order_key]}{summary_suffix}]"
+            processed_parts.extend((chunk_prefix, f"{chunk[section_content_key]}"))
+            included_chunks.append(chunk)
+
+        if in_skip:
+            processed_parts.append(f"[... {skip_char_count} characters skipped ...]")
+
+        if reverse:
+            processed_parts = list(reversed(processed_parts))
+
+        return processed_parts
+
+    def render_hierarchy_headers(
+        self,
+        current_chunk: Dict,
+        chunks: List[Dict],
+        doc_header_key: str,
+    ) -> str:
+        """
+        Render headers for the current chunk's hierarchy.
+
+        Args:
+            current_chunk (Dict): The current chunk being processed.
+            chunks (List[Dict]): List of chunks up to and including the current chunk.
+            doc_header_key (str): The key for the headers in the current chunk.
+        Returns:
+            str: Renderted headers in the current chunk's hierarchy.
+        """
+        current_hierarchy = {}
+
+        if doc_header_key is None:
+            return ""
+
+        # Find the largest/highest level in the current chunk
+        current_chunk_headers = current_chunk.get(doc_header_key, [])
+        highest_level = float("inf")  # Initialize with positive infinity
+        for header_info in current_chunk_headers:
+            level = header_info.get("level")
+            if level is not None and level < highest_level:
+                highest_level = level
+
+        # If no headers found in the current chunk, set highest_level to None
+        if highest_level == float("inf"):
+            highest_level = None
+
+        for chunk in chunks:
+            for header_info in chunk.get(doc_header_key, []):
+                header = header_info["header"]
+                level = header_info["level"]
+                if header and level:
+                    current_hierarchy[level] = header
+                    # Clear lower levels when a higher level header is found
+                    for lower_level in range(level + 1, len(current_hierarchy) + 1):
+                        if lower_level in current_hierarchy:
+                            current_hierarchy[lower_level] = None
+
+        rendered_headers = [
+            f"{'#' * level} {header}"
+            for level, header in sorted(current_hierarchy.items())
+            if header is not None and (highest_level is None or level < highest_level)
+        ]
+        rendered_headers = " > ".join(rendered_headers)
+        return f"_Current Section:_ {rendered_headers}" if rendered_headers else ""
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ __init__(*args, **kwargs) + +

+ + +
+ +

Initialize the GatherOperation.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ *args + + Any + +
+

Variable length argument list.

+
+
+ () +
+ **kwargs + + Any + +
+

Arbitrary keyword arguments.

+
+
+ {} +
+ +
+ Source code in docetl/operations/gather.py +
26
+27
+28
+29
+30
+31
+32
+33
+34
def __init__(self, *args: Any, **kwargs: Any) -> None:
+    """
+    Initialize the GatherOperation.
+
+    Args:
+        *args: Variable length argument list.
+        **kwargs: Arbitrary keyword arguments.
+    """
+    super().__init__(*args, **kwargs)
+
+
+
+ +
+ +
+ + +

+ execute(input_data) + +

+ + +
+ +

Execute the gather operation on the input data.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ input_data + + List[Dict] + +
+

The input data to process.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ Tuple[List[Dict], float] + +
+

Tuple[List[Dict], float]: A tuple containing the processed results and the cost of the operation.

+
+
+ +
+ Source code in docetl/operations/gather.py +
 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+    """
+    Execute the gather operation on the input data.
+
+    Args:
+        input_data (List[Dict]): The input data to process.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the processed results and the cost of the operation.
+    """
+    content_key = self.config["content_key"]
+    doc_id_key = self.config["doc_id_key"]
+    order_key = self.config["order_key"]
+    peripheral_config = self.config.get("peripheral_chunks", {})
+    main_chunk_start = self.config.get(
+        "main_chunk_start", "--- Begin Main Chunk ---"
+    )
+    main_chunk_end = self.config.get("main_chunk_end", "--- End Main Chunk ---")
+    doc_header_key = self.config.get("doc_header_key", None)
+    results = []
+    cost = 0.0
+
+    # Group chunks by document ID
+    grouped_chunks = {}
+    for item in input_data:
+        doc_id = item[doc_id_key]
+        if doc_id not in grouped_chunks:
+            grouped_chunks[doc_id] = []
+        grouped_chunks[doc_id].append(item)
+
+    # Process each group of chunks
+    for chunks in grouped_chunks.values():
+        # Sort chunks by their order within the document
+        chunks.sort(key=lambda x: x[order_key])
+
+        # Process each chunk with its peripheral context and headers
+        for i, chunk in enumerate(chunks):
+            rendered_chunk = self.render_chunk_with_context(
+                chunks,
+                i,
+                peripheral_config,
+                content_key,
+                order_key,
+                main_chunk_start,
+                main_chunk_end,
+                doc_header_key,
+            )
+
+            result = chunk.copy()
+            result[f"{content_key}_rendered"] = rendered_chunk
+            results.append(result)
+
+    return results, cost
+
+
+
+ +
+ +
+ + +

+ process_peripheral_chunks(chunks, config, content_key, order_key, reverse=False) + +

+ + +
+ +

Process peripheral chunks according to the configuration.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ chunks + + List[Dict] + +
+

List of chunks to process.

+
+
+ required +
+ config + + Dict + +
+

Configuration for processing peripheral chunks.

+
+
+ required +
+ content_key + + str + +
+

Key for the content in each chunk.

+
+
+ required +
+ order_key + + str + +
+

Key for the order of each chunk.

+
+
+ required +
+ reverse + + bool + +
+

Whether to process chunks in reverse order. Defaults to False.

+
+
+ False +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ List[str] + +
+

List[str]: List of processed chunk strings.

+
+
+ +
+ Source code in docetl/operations/gather.py +
191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
def process_peripheral_chunks(
+    self,
+    chunks: List[Dict],
+    config: Dict,
+    content_key: str,
+    order_key: str,
+    reverse: bool = False,
+) -> List[str]:
+    """
+    Process peripheral chunks according to the configuration.
+
+    Args:
+        chunks (List[Dict]): List of chunks to process.
+        config (Dict): Configuration for processing peripheral chunks.
+        content_key (str): Key for the content in each chunk.
+        order_key (str): Key for the order of each chunk.
+        reverse (bool, optional): Whether to process chunks in reverse order. Defaults to False.
+
+    Returns:
+        List[str]: List of processed chunk strings.
+    """
+    if reverse:
+        chunks = list(reversed(chunks))
+
+    processed_parts = []
+    included_chunks = []
+    total_chunks = len(chunks)
+
+    head_config = config.get("head", {})
+    tail_config = config.get("tail", {})
+
+    head_count = int(head_config.get("count", 0))
+    tail_count = int(tail_config.get("count", 0))
+    in_skip = False
+    skip_char_count = 0
+
+    for i, chunk in enumerate(chunks):
+        if i < head_count:
+            section = "head"
+        elif i >= total_chunks - tail_count:
+            section = "tail"
+        elif "middle" in config:
+            section = "middle"
+        else:
+            # Show number of characters skipped
+            skipped_chars = len(chunk[content_key])
+            if not in_skip:
+                skip_char_count = skipped_chars
+                in_skip = True
+            else:
+                skip_char_count += skipped_chars
+
+            continue
+
+        if in_skip:
+            processed_parts.append(
+                f"[... {skip_char_count} characters skipped ...]"
+            )
+            in_skip = False
+            skip_char_count = 0
+
+        section_config = config.get(section, {})
+        section_content_key = section_config.get("content_key", content_key)
+
+        is_summary = section_content_key != content_key
+        summary_suffix = " (Summary)" if is_summary else ""
+
+        chunk_prefix = f"[Chunk {chunk[order_key]}{summary_suffix}]"
+        processed_parts.extend((chunk_prefix, f"{chunk[section_content_key]}"))
+        included_chunks.append(chunk)
+
+    if in_skip:
+        processed_parts.append(f"[... {skip_char_count} characters skipped ...]")
+
+    if reverse:
+        processed_parts = list(reversed(processed_parts))
+
+    return processed_parts
+
+
+
+ +
+ +
+ + +

+ render_chunk_with_context(chunks, current_index, peripheral_config, content_key, order_key, main_chunk_start, main_chunk_end, doc_header_key) + +

+ + +
+ +

Render a chunk with its peripheral context and headers.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ chunks + + List[Dict] + +
+

List of all chunks in the document.

+
+
+ required +
+ current_index + + int + +
+

Index of the current chunk being processed.

+
+
+ required +
+ peripheral_config + + Dict + +
+

Configuration for peripheral chunks.

+
+
+ required +
+ content_key + + str + +
+

Key for the content in each chunk.

+
+
+ required +
+ order_key + + str + +
+

Key for the order of each chunk.

+
+
+ required +
+ main_chunk_start + + str + +
+

String to mark the start of the main chunk.

+
+
+ required +
+ main_chunk_end + + str + +
+

String to mark the end of the main chunk.

+
+
+ required +
+ doc_header_key + + str + +
+

The key for the headers in the current chunk.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
Name TypeDescription
str + str + +
+

Renderted chunk with context and headers.

+
+
+ +
+ Source code in docetl/operations/gather.py +
126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
def render_chunk_with_context(
+    self,
+    chunks: List[Dict],
+    current_index: int,
+    peripheral_config: Dict,
+    content_key: str,
+    order_key: str,
+    main_chunk_start: str,
+    main_chunk_end: str,
+    doc_header_key: str,
+) -> str:
+    """
+    Render a chunk with its peripheral context and headers.
+
+    Args:
+        chunks (List[Dict]): List of all chunks in the document.
+        current_index (int): Index of the current chunk being processed.
+        peripheral_config (Dict): Configuration for peripheral chunks.
+        content_key (str): Key for the content in each chunk.
+        order_key (str): Key for the order of each chunk.
+        main_chunk_start (str): String to mark the start of the main chunk.
+        main_chunk_end (str): String to mark the end of the main chunk.
+        doc_header_key (str): The key for the headers in the current chunk.
+
+    Returns:
+        str: Renderted chunk with context and headers.
+    """
+    combined_parts = ["--- Previous Context ---"]
+
+    combined_parts.extend(
+        self.process_peripheral_chunks(
+            chunks[:current_index],
+            peripheral_config.get("previous", {}),
+            content_key,
+            order_key,
+        )
+    )
+    combined_parts.append("--- End Previous Context ---\n")
+
+    # Process main chunk
+    main_chunk = chunks[current_index]
+    if headers := self.render_hierarchy_headers(
+        main_chunk, chunks[: current_index + 1], doc_header_key
+    ):
+        combined_parts.append(headers)
+    combined_parts.extend(
+        (
+            f"{main_chunk_start}",
+            f"{main_chunk[content_key]}",
+            f"{main_chunk_end}",
+            "\n--- Next Context ---",
+        )
+    )
+    combined_parts.extend(
+        self.process_peripheral_chunks(
+            chunks[current_index + 1 :],
+            peripheral_config.get("next", {}),
+            content_key,
+            order_key,
+        )
+    )
+    combined_parts.append("--- End Next Context ---")
+
+    return "\n".join(combined_parts)
+
+
+
+ +
+ +
+ + +

+ render_hierarchy_headers(current_chunk, chunks, doc_header_key) + +

+ + +
+ +

Render headers for the current chunk's hierarchy.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ current_chunk + + Dict + +
+

The current chunk being processed.

+
+
+ required +
+ chunks + + List[Dict] + +
+

List of chunks up to and including the current chunk.

+
+
+ required +
+ doc_header_key + + str + +
+

The key for the headers in the current chunk.

+
+
+ required +
+

Returns: + str: Renderted headers in the current chunk's hierarchy.

+ +
+ Source code in docetl/operations/gather.py +
270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
def render_hierarchy_headers(
+    self,
+    current_chunk: Dict,
+    chunks: List[Dict],
+    doc_header_key: str,
+) -> str:
+    """
+    Render headers for the current chunk's hierarchy.
+
+    Args:
+        current_chunk (Dict): The current chunk being processed.
+        chunks (List[Dict]): List of chunks up to and including the current chunk.
+        doc_header_key (str): The key for the headers in the current chunk.
+    Returns:
+        str: Renderted headers in the current chunk's hierarchy.
+    """
+    current_hierarchy = {}
+
+    if doc_header_key is None:
+        return ""
+
+    # Find the largest/highest level in the current chunk
+    current_chunk_headers = current_chunk.get(doc_header_key, [])
+    highest_level = float("inf")  # Initialize with positive infinity
+    for header_info in current_chunk_headers:
+        level = header_info.get("level")
+        if level is not None and level < highest_level:
+            highest_level = level
+
+    # If no headers found in the current chunk, set highest_level to None
+    if highest_level == float("inf"):
+        highest_level = None
+
+    for chunk in chunks:
+        for header_info in chunk.get(doc_header_key, []):
+            header = header_info["header"]
+            level = header_info["level"]
+            if header and level:
+                current_hierarchy[level] = header
+                # Clear lower levels when a higher level header is found
+                for lower_level in range(level + 1, len(current_hierarchy) + 1):
+                    if lower_level in current_hierarchy:
+                        current_hierarchy[lower_level] = None
+
+    rendered_headers = [
+        f"{'#' * level} {header}"
+        for level, header in sorted(current_hierarchy.items())
+        if header is not None and (highest_level is None or level < highest_level)
+    ]
+    rendered_headers = " > ".join(rendered_headers)
+    return f"_Current Section:_ {rendered_headers}" if rendered_headers else ""
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Perform a syntax check on the operation configuration.

+ + +

Raises:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If required keys are missing or if there are configuration errors.

+
+
+ TypeError + +
+

If main_chunk_start or main_chunk_end are not strings.

+
+
+ +
+ Source code in docetl/operations/gather.py +
36
+37
+38
+39
+40
+41
+42
+43
+44
+45
+46
+47
+48
+49
+50
+51
+52
+53
+54
+55
+56
+57
+58
+59
+60
+61
+62
+63
+64
+65
+66
+67
+68
+69
+70
def syntax_check(self) -> None:
+    """
+    Perform a syntax check on the operation configuration.
+
+    Raises:
+        ValueError: If required keys are missing or if there are configuration errors.
+        TypeError: If main_chunk_start or main_chunk_end are not strings.
+    """
+    required_keys = ["content_key", "doc_id_key", "order_key"]
+    for key in required_keys:
+        if key not in self.config:
+            raise ValueError(
+                f"Missing required key '{key}' in GatherOperation configuration"
+            )
+
+    peripheral_config = self.config.get("peripheral_chunks", {})
+    for direction in ["previous", "next"]:
+        if direction not in peripheral_config:
+            continue
+        for section in ["head", "middle", "tail"]:
+            if section in peripheral_config[direction]:
+                section_config = peripheral_config[direction][section]
+                if section != "middle" and "count" not in section_config:
+                    raise ValueError(
+                        f"Missing 'count' in {direction}.{section} configuration"
+                    )
+
+    if "main_chunk_start" in self.config and not isinstance(
+        self.config["main_chunk_start"], str
+    ):
+        raise TypeError("'main_chunk_start' must be a string")
+    if "main_chunk_end" in self.config and not isinstance(
+        self.config["main_chunk_end"], str
+    ):
+        raise TypeError("'main_chunk_end' must be a string")
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.operations.unnest.UnnestOperation + + +

+ + +
+

+ Bases: BaseOperation

+ + +

A class that represents an operation to unnest a list-like or dictionary value in a dictionary into multiple dictionaries.

+

This operation takes a list of dictionaries and a specified key, and creates new dictionaries based on the value type: +- For list-like values: Creates a new dictionary for each element in the list, copying all other key-value pairs. +- For dictionary values: Expands specified fields from the nested dictionary into the parent dictionary.

+ + +
+ Inherits from +

BaseOperation

+

Usage: +

from docetl.operations import UnnestOperation
+
+# Unnesting a list
+config_list = {"unnest_key": "tags"}
+input_data_list = [
+    {"id": 1, "tags": ["a", "b", "c"]},
+    {"id": 2, "tags": ["d", "e"]}
+]
+
+unnest_op_list = UnnestOperation(config_list)
+result_list, _ = unnest_op_list.execute(input_data_list)
+
+# Result will be:
+# [
+#     {"id": 1, "tags": "a"},
+#     {"id": 1, "tags": "b"},
+#     {"id": 1, "tags": "c"},
+#     {"id": 2, "tags": "d"},
+#     {"id": 2, "tags": "e"}
+# ]
+
+# Unnesting a dictionary
+config_dict = {"unnest_key": "user", "expand_fields": ["name", "age"]}
+input_data_dict = [
+    {"id": 1, "user": {"name": "Alice", "age": 30, "email": "alice@example.com"}},
+    {"id": 2, "user": {"name": "Bob", "age": 25, "email": "bob@example.com"}}
+]
+
+unnest_op_dict = UnnestOperation(config_dict)
+result_dict, _ = unnest_op_dict.execute(input_data_dict)
+
+# Result will be:
+# [
+#     {"id": 1, "name": "Alice", "age": 30, "user": {"name": "Alice", "age": 30, "email": "alice@example.com"}},
+#     {"id": 2, "name": "Bob", "age": 25, "user": {"name": "Bob", "age": 25, "email": "bob@example.com"}}
+# ]
+

+ + + + + + +
+ Source code in docetl/operations/unnest.py +
  7
+  8
+  9
+ 10
+ 11
+ 12
+ 13
+ 14
+ 15
+ 16
+ 17
+ 18
+ 19
+ 20
+ 21
+ 22
+ 23
+ 24
+ 25
+ 26
+ 27
+ 28
+ 29
+ 30
+ 31
+ 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
class UnnestOperation(BaseOperation):
+    """
+    A class that represents an operation to unnest a list-like or dictionary value in a dictionary into multiple dictionaries.
+
+    This operation takes a list of dictionaries and a specified key, and creates new dictionaries based on the value type:
+    - For list-like values: Creates a new dictionary for each element in the list, copying all other key-value pairs.
+    - For dictionary values: Expands specified fields from the nested dictionary into the parent dictionary.
+
+    Inherits from:
+        BaseOperation
+
+    Usage:
+    ```python
+    from docetl.operations import UnnestOperation
+
+    # Unnesting a list
+    config_list = {"unnest_key": "tags"}
+    input_data_list = [
+        {"id": 1, "tags": ["a", "b", "c"]},
+        {"id": 2, "tags": ["d", "e"]}
+    ]
+
+    unnest_op_list = UnnestOperation(config_list)
+    result_list, _ = unnest_op_list.execute(input_data_list)
+
+    # Result will be:
+    # [
+    #     {"id": 1, "tags": "a"},
+    #     {"id": 1, "tags": "b"},
+    #     {"id": 1, "tags": "c"},
+    #     {"id": 2, "tags": "d"},
+    #     {"id": 2, "tags": "e"}
+    # ]
+
+    # Unnesting a dictionary
+    config_dict = {"unnest_key": "user", "expand_fields": ["name", "age"]}
+    input_data_dict = [
+        {"id": 1, "user": {"name": "Alice", "age": 30, "email": "alice@example.com"}},
+        {"id": 2, "user": {"name": "Bob", "age": 25, "email": "bob@example.com"}}
+    ]
+
+    unnest_op_dict = UnnestOperation(config_dict)
+    result_dict, _ = unnest_op_dict.execute(input_data_dict)
+
+    # Result will be:
+    # [
+    #     {"id": 1, "name": "Alice", "age": 30, "user": {"name": "Alice", "age": 30, "email": "alice@example.com"}},
+    #     {"id": 2, "name": "Bob", "age": 25, "user": {"name": "Bob", "age": 25, "email": "bob@example.com"}}
+    # ]
+    ```
+    """
+
+    class schema(BaseOperation.schema):
+        type: str = "unnest"
+        unnest_key: str
+        keep_empty: Optional[bool] = None
+        expand_fields: Optional[List[str]] = None
+        recursive: Optional[bool] = None
+        depth: Optional[int] = None
+
+    def syntax_check(self) -> None:
+        """
+        Checks if the required configuration key is present in the operation's config.
+
+        Raises:
+            ValueError: If the required 'unnest_key' is missing from the configuration.
+        """
+
+        required_keys = ["unnest_key"]
+        for key in required_keys:
+            if key not in self.config:
+                raise ValueError(
+                    f"Missing required key '{key}' in UnnestOperation configuration"
+                )
+
+    def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+        """
+        Executes the unnest operation on the input data.
+
+        Args:
+            input_data (List[Dict]): A list of dictionaries to process.
+
+        Returns:
+            Tuple[List[Dict], float]: A tuple containing the processed list of dictionaries
+            and a float value (always 0 in this implementation).
+
+        Raises:
+            KeyError: If the specified unnest_key is not found in an input dictionary.
+            TypeError: If the value of the unnest_key is not iterable (list, tuple, set, or dict).
+            ValueError: If unnesting a dictionary and 'expand_fields' is not provided in the config.
+
+        The operation supports unnesting of both list-like values and dictionary values:
+
+        1. For list-like values (list, tuple, set):
+           Each element in the list becomes a separate dictionary in the output.
+
+        2. For dictionary values:
+           The operation expands specified fields from the nested dictionary into the parent dictionary.
+           The 'expand_fields' config parameter must be provided to specify which fields to expand.
+
+        Examples:
+        ```python
+        # Unnesting a list
+        unnest_op = UnnestOperation({"unnest_key": "colors"})
+        input_data = [
+            {"id": 1, "colors": ["red", "blue"]},
+            {"id": 2, "colors": ["green"]}
+        ]
+        result, _ = unnest_op.execute(input_data)
+        # Result will be:
+        # [
+        #     {"id": 1, "colors": "red"},
+        #     {"id": 1, "colors": "blue"},
+        #     {"id": 2, "colors": "green"}
+        # ]
+
+        # Unnesting a dictionary
+        unnest_op = UnnestOperation({"unnest_key": "details", "expand_fields": ["color", "size"]})
+        input_data = [
+            {"id": 1, "details": {"color": "red", "size": "large", "stock": 5}},
+            {"id": 2, "details": {"color": "blue", "size": "medium", "stock": 3}}
+        ]
+        result, _ = unnest_op.execute(input_data)
+        # Result will be:
+        # [
+        #     {"id": 1, "details": {"color": "red", "size": "large", "stock": 5}, "color": "red", "size": "large"},
+        #     {"id": 2, "details": {"color": "blue", "size": "medium", "stock": 3}, "color": "blue", "size": "medium"}
+        # ]
+        ```
+
+        Note: When unnesting dictionaries, the original nested dictionary is preserved in the output,
+        and the specified fields are expanded into the parent dictionary.
+        """
+
+        unnest_key = self.config["unnest_key"]
+        recursive = self.config.get("recursive", False)
+        depth = self.config.get("depth", None)
+        if not depth:
+            depth = 1 if not recursive else float("inf")
+        results = []
+
+        def unnest_recursive(item, key, level=0):
+            if level == 0 and not isinstance(item[key], (list, tuple, set, dict)):
+                raise TypeError(f"Value of unnest key '{key}' is not iterable")
+
+            if level > 0 and not isinstance(item[key], (list, tuple, set, dict)):
+                return [item]
+
+            if level >= depth:
+                return [item]
+
+            if isinstance(item[key], dict):
+                expand_fields = self.config.get("expand_fields")
+                if expand_fields is None:
+                    expand_fields = item[key].keys()
+                new_item = copy.deepcopy(item)
+                for field in expand_fields:
+                    if field in new_item[key]:
+                        new_item[field] = new_item[key][field]
+                    else:
+                        new_item[field] = None
+                return [new_item]
+            else:
+                nested_results = []
+                for value in item[key]:
+                    new_item = copy.deepcopy(item)
+                    new_item[key] = value
+                    if recursive and isinstance(value, (list, tuple, set, dict)):
+                        nested_results.extend(
+                            unnest_recursive(new_item, key, level + 1)
+                        )
+                    else:
+                        nested_results.append(new_item)
+                return nested_results
+
+        for item in input_data:
+            if unnest_key not in item:
+                raise KeyError(
+                    f"Unnest key '{unnest_key}' not found in item. Other keys are {item.keys()}"
+                )
+
+            results.extend(unnest_recursive(item, unnest_key))
+
+            if not item[unnest_key] and self.config.get("keep_empty", False):
+                expand_fields = self.config.get("expand_fields")
+                new_item = copy.deepcopy(item)
+                if isinstance(item[unnest_key], dict):
+                    if expand_fields is None:
+                        expand_fields = item[unnest_key].keys()
+                    for field in expand_fields:
+                        new_item[field] = None
+                else:
+                    new_item[unnest_key] = None
+                results.append(new_item)
+
+        # Assert that no keys are missing after the operation
+        if results:
+            original_keys = set(input_data[0].keys())
+            assert original_keys.issubset(
+                set(results[0].keys())
+            ), "Keys lost during unnest operation"
+
+        return results, 0
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ execute(input_data) + +

+ + +
+ +

Executes the unnest operation on the input data.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ input_data + + List[Dict] + +
+

A list of dictionaries to process.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ List[Dict] + +
+

Tuple[List[Dict], float]: A tuple containing the processed list of dictionaries

+
+
+ float + +
+

and a float value (always 0 in this implementation).

+
+
+ + +

Raises:

+ + + + + + + + + + + + + + + + + + + + + +
TypeDescription
+ KeyError + +
+

If the specified unnest_key is not found in an input dictionary.

+
+
+ TypeError + +
+

If the value of the unnest_key is not iterable (list, tuple, set, or dict).

+
+
+ ValueError + +
+

If unnesting a dictionary and 'expand_fields' is not provided in the config.

+
+
+

The operation supports unnesting of both list-like values and dictionary values:

+
    +
  1. +

    For list-like values (list, tuple, set): + Each element in the list becomes a separate dictionary in the output.

    +
  2. +
  3. +

    For dictionary values: + The operation expands specified fields from the nested dictionary into the parent dictionary. + The 'expand_fields' config parameter must be provided to specify which fields to expand.

    +
  4. +
+

Examples: +

# Unnesting a list
+unnest_op = UnnestOperation({"unnest_key": "colors"})
+input_data = [
+    {"id": 1, "colors": ["red", "blue"]},
+    {"id": 2, "colors": ["green"]}
+]
+result, _ = unnest_op.execute(input_data)
+# Result will be:
+# [
+#     {"id": 1, "colors": "red"},
+#     {"id": 1, "colors": "blue"},
+#     {"id": 2, "colors": "green"}
+# ]
+
+# Unnesting a dictionary
+unnest_op = UnnestOperation({"unnest_key": "details", "expand_fields": ["color", "size"]})
+input_data = [
+    {"id": 1, "details": {"color": "red", "size": "large", "stock": 5}},
+    {"id": 2, "details": {"color": "blue", "size": "medium", "stock": 3}}
+]
+result, _ = unnest_op.execute(input_data)
+# Result will be:
+# [
+#     {"id": 1, "details": {"color": "red", "size": "large", "stock": 5}, "color": "red", "size": "large"},
+#     {"id": 2, "details": {"color": "blue", "size": "medium", "stock": 3}, "color": "blue", "size": "medium"}
+# ]
+

+

Note: When unnesting dictionaries, the original nested dictionary is preserved in the output, +and the specified fields are expanded into the parent dictionary.

+ +
+ Source code in docetl/operations/unnest.py +
 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
def execute(self, input_data: List[Dict]) -> Tuple[List[Dict], float]:
+    """
+    Executes the unnest operation on the input data.
+
+    Args:
+        input_data (List[Dict]): A list of dictionaries to process.
+
+    Returns:
+        Tuple[List[Dict], float]: A tuple containing the processed list of dictionaries
+        and a float value (always 0 in this implementation).
+
+    Raises:
+        KeyError: If the specified unnest_key is not found in an input dictionary.
+        TypeError: If the value of the unnest_key is not iterable (list, tuple, set, or dict).
+        ValueError: If unnesting a dictionary and 'expand_fields' is not provided in the config.
+
+    The operation supports unnesting of both list-like values and dictionary values:
+
+    1. For list-like values (list, tuple, set):
+       Each element in the list becomes a separate dictionary in the output.
+
+    2. For dictionary values:
+       The operation expands specified fields from the nested dictionary into the parent dictionary.
+       The 'expand_fields' config parameter must be provided to specify which fields to expand.
+
+    Examples:
+    ```python
+    # Unnesting a list
+    unnest_op = UnnestOperation({"unnest_key": "colors"})
+    input_data = [
+        {"id": 1, "colors": ["red", "blue"]},
+        {"id": 2, "colors": ["green"]}
+    ]
+    result, _ = unnest_op.execute(input_data)
+    # Result will be:
+    # [
+    #     {"id": 1, "colors": "red"},
+    #     {"id": 1, "colors": "blue"},
+    #     {"id": 2, "colors": "green"}
+    # ]
+
+    # Unnesting a dictionary
+    unnest_op = UnnestOperation({"unnest_key": "details", "expand_fields": ["color", "size"]})
+    input_data = [
+        {"id": 1, "details": {"color": "red", "size": "large", "stock": 5}},
+        {"id": 2, "details": {"color": "blue", "size": "medium", "stock": 3}}
+    ]
+    result, _ = unnest_op.execute(input_data)
+    # Result will be:
+    # [
+    #     {"id": 1, "details": {"color": "red", "size": "large", "stock": 5}, "color": "red", "size": "large"},
+    #     {"id": 2, "details": {"color": "blue", "size": "medium", "stock": 3}, "color": "blue", "size": "medium"}
+    # ]
+    ```
+
+    Note: When unnesting dictionaries, the original nested dictionary is preserved in the output,
+    and the specified fields are expanded into the parent dictionary.
+    """
+
+    unnest_key = self.config["unnest_key"]
+    recursive = self.config.get("recursive", False)
+    depth = self.config.get("depth", None)
+    if not depth:
+        depth = 1 if not recursive else float("inf")
+    results = []
+
+    def unnest_recursive(item, key, level=0):
+        if level == 0 and not isinstance(item[key], (list, tuple, set, dict)):
+            raise TypeError(f"Value of unnest key '{key}' is not iterable")
+
+        if level > 0 and not isinstance(item[key], (list, tuple, set, dict)):
+            return [item]
+
+        if level >= depth:
+            return [item]
+
+        if isinstance(item[key], dict):
+            expand_fields = self.config.get("expand_fields")
+            if expand_fields is None:
+                expand_fields = item[key].keys()
+            new_item = copy.deepcopy(item)
+            for field in expand_fields:
+                if field in new_item[key]:
+                    new_item[field] = new_item[key][field]
+                else:
+                    new_item[field] = None
+            return [new_item]
+        else:
+            nested_results = []
+            for value in item[key]:
+                new_item = copy.deepcopy(item)
+                new_item[key] = value
+                if recursive and isinstance(value, (list, tuple, set, dict)):
+                    nested_results.extend(
+                        unnest_recursive(new_item, key, level + 1)
+                    )
+                else:
+                    nested_results.append(new_item)
+            return nested_results
+
+    for item in input_data:
+        if unnest_key not in item:
+            raise KeyError(
+                f"Unnest key '{unnest_key}' not found in item. Other keys are {item.keys()}"
+            )
+
+        results.extend(unnest_recursive(item, unnest_key))
+
+        if not item[unnest_key] and self.config.get("keep_empty", False):
+            expand_fields = self.config.get("expand_fields")
+            new_item = copy.deepcopy(item)
+            if isinstance(item[unnest_key], dict):
+                if expand_fields is None:
+                    expand_fields = item[unnest_key].keys()
+                for field in expand_fields:
+                    new_item[field] = None
+            else:
+                new_item[unnest_key] = None
+            results.append(new_item)
+
+    # Assert that no keys are missing after the operation
+    if results:
+        original_keys = set(input_data[0].keys())
+        assert original_keys.issubset(
+            set(results[0].keys())
+        ), "Keys lost during unnest operation"
+
+    return results, 0
+
+
+
+ +
+ +
+ + +

+ syntax_check() + +

+ + +
+ +

Checks if the required configuration key is present in the operation's config.

+ + +

Raises:

+ + + + + + + + + + + + + +
TypeDescription
+ ValueError + +
+

If the required 'unnest_key' is missing from the configuration.

+
+
+ +
+ Source code in docetl/operations/unnest.py +
67
+68
+69
+70
+71
+72
+73
+74
+75
+76
+77
+78
+79
+80
def syntax_check(self) -> None:
+    """
+    Checks if the required configuration key is present in the operation's config.
+
+    Raises:
+        ValueError: If the required 'unnest_key' is missing from the configuration.
+    """
+
+    required_keys = ["unnest_key"]
+    for key in required_keys:
+        if key not in self.config:
+            raise ValueError(
+                f"Missing required key '{key}' in UnnestOperation configuration"
+            )
+
+
+
+ +
+ + + +
+ +
+ +
+ + + + + + + + + + + + + +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/api-reference/optimizers/index.html b/api-reference/optimizers/index.html new file mode 100644 index 00000000..6bd708f9 --- /dev/null +++ b/api-reference/optimizers/index.html @@ -0,0 +1,11123 @@ + + + + + + + + + + + + + + + + + + + + + + + docetl.optimizers - docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + + + + +

docetl.optimizers

+ +
+ + + +

+ docetl.optimizers.map_optimizer.optimizer.MapOptimizer + + +

+ + +
+ + +

A class for optimizing map operations in data processing pipelines.

+

This optimizer analyzes the input operation configuration and data, +and generates optimized plans for executing the operation. It can +create plans for chunking, metadata extraction, gleaning, chain +decomposition, and parallel execution.

+ + +

Attributes:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescription
config + Dict[str, Any] + +
+

The configuration dictionary for the optimizer.

+
+
console + Console + +
+

A Rich console object for pretty printing.

+
+
llm_client + LLMClient + +
+

A client for interacting with a language model.

+
+
_run_operation + Callable + +
+

A function to execute operations.

+
+
max_threads + int + +
+

The maximum number of threads to use for parallel execution.

+
+
timeout + int + +
+

The timeout in seconds for operation execution.

+
+
+ + + + + + +
+ Source code in docetl/optimizers/map_optimizer/optimizer.py +
 21
+ 22
+ 23
+ 24
+ 25
+ 26
+ 27
+ 28
+ 29
+ 30
+ 31
+ 32
+ 33
+ 34
+ 35
+ 36
+ 37
+ 38
+ 39
+ 40
+ 41
+ 42
+ 43
+ 44
+ 45
+ 46
+ 47
+ 48
+ 49
+ 50
+ 51
+ 52
+ 53
+ 54
+ 55
+ 56
+ 57
+ 58
+ 59
+ 60
+ 61
+ 62
+ 63
+ 64
+ 65
+ 66
+ 67
+ 68
+ 69
+ 70
+ 71
+ 72
+ 73
+ 74
+ 75
+ 76
+ 77
+ 78
+ 79
+ 80
+ 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
class MapOptimizer:
+    """
+    A class for optimizing map operations in data processing pipelines.
+
+    This optimizer analyzes the input operation configuration and data,
+    and generates optimized plans for executing the operation. It can
+    create plans for chunking, metadata extraction, gleaning, chain
+    decomposition, and parallel execution.
+
+    Attributes:
+        config (Dict[str, Any]): The configuration dictionary for the optimizer.
+        console (Console): A Rich console object for pretty printing.
+        llm_client (LLMClient): A client for interacting with a language model.
+        _run_operation (Callable): A function to execute operations.
+        max_threads (int): The maximum number of threads to use for parallel execution.
+        timeout (int): The timeout in seconds for operation execution.
+
+    """
+
+    def __init__(
+        self,
+        runner,
+        config: Dict[str, Any],
+        console: Console,
+        llm_client: LLMClient,
+        max_threads: int,
+        run_operation: Callable,
+        timeout: int = 10,
+        is_filter: bool = False,
+    ):
+        """
+        Initialize the MapOptimizer.
+
+        Args:
+            config (Dict[str, Any]): The configuration dictionary for the optimizer.
+            console (Console): A Rich console object for pretty printing.
+            llm_client (LLMClient): A client for interacting with a language model.
+            max_threads (int): The maximum number of threads to use for parallel execution.
+            run_operation (Callable): A function to execute operations.
+            timeout (int, optional): The timeout in seconds for operation execution. Defaults to 10.
+            is_filter (bool, optional): If True, the operation is a filter operation. Defaults to False.
+        """
+        self.runner = runner
+        self.config = config
+        self.console = console
+        self.llm_client = llm_client
+        self._run_operation = run_operation
+        self.max_threads = max_threads
+        self.timeout = timeout
+        self._num_plans_to_evaluate_in_parallel = 5
+        self.is_filter = is_filter
+        self.k_to_pairwise_compare = 6
+
+        self.plan_generator = PlanGenerator(
+            runner, llm_client, console, config, run_operation, max_threads, is_filter
+        )
+        self.evaluator = Evaluator(
+            llm_client,
+            console,
+            run_operation,
+            timeout,
+            self._num_plans_to_evaluate_in_parallel,
+            is_filter,
+        )
+        self.prompt_generator = PromptGenerator(
+            runner, llm_client, console, config, max_threads, is_filter
+        )
+
+    def should_optimize(self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]) -> Tuple[str, List[Dict[str, Any]], List[Dict[str, Any]]]:
+        """
+        Determine if the given operation configuration should be optimized.
+        """
+        input_data, output_data, _, _, validator_prompt, assessment, data_exceeds_limit = self._should_optimize_helper(op_config, input_data)
+        if data_exceeds_limit or assessment.get("needs_improvement", True):
+            assessment_str = "\n".join(assessment.get("reasons", [])) + "\n\nHere are some improvements that may help:\n" + "\n".join(assessment.get("improvements", []))
+            if data_exceeds_limit:
+                assessment_str += "\nAlso, the input data exceeds the token limit."
+            return assessment_str, input_data, output_data
+        else:
+            return "", input_data, output_data
+
+
+    def _should_optimize_helper(self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]], int, float, str, Dict[str, Any], bool]:
+        """
+        Determine if the given operation configuration should be optimized.
+        Create a custom validator prompt and assess the operation's performance
+        using the validator.
+        """
+        self.console.post_optimizer_status(StageType.SAMPLE_RUN)
+        input_data = copy.deepcopy(input_data)
+        # Add id to each input_data
+        for i in range(len(input_data)):
+            input_data[i]["_map_opt_id"] = str(uuid.uuid4())
+
+        # Define the token limit (adjust as needed)
+        model_input_context_length = model_cost.get(
+            op_config.get("model", self.config.get("default_model")), {}
+        ).get("max_input_tokens", 8192)
+
+        # Render the prompt with all sample inputs and count tokens
+        total_tokens = 0
+        exceed_count = 0
+        for sample in input_data:
+            rendered_prompt = Template(op_config["prompt"]).render(input=sample)
+            prompt_tokens = count_tokens(
+                rendered_prompt,
+                op_config.get("model", self.config.get("default_model")),
+            )
+            total_tokens += prompt_tokens
+
+            if prompt_tokens > model_input_context_length:
+                exceed_count += 1
+
+        # Calculate average tokens and percentage of samples exceeding limit
+        avg_tokens = total_tokens / len(input_data)
+        exceed_percentage = (exceed_count / len(input_data)) * 100
+
+        data_exceeds_limit = exceed_count > 0
+        if exceed_count > 0:
+            self.console.log(
+                f"[yellow]Warning: {exceed_percentage:.2f}% of prompts exceed token limit. "
+                f"Average token count: {avg_tokens:.2f}. "
+                f"Truncating input data when generating validators.[/yellow]"
+            )
+
+        # Execute the original operation on the sample data
+        no_change_start = time.time()
+        output_data = self._run_operation(op_config, input_data, is_build=True)
+        no_change_runtime = time.time() - no_change_start
+
+        # Capture output for the sample run
+        self.runner.captured_output.save_optimizer_output(
+            stage_type=StageType.SAMPLE_RUN,
+            output={
+                "operation_config": op_config,
+                "input_data": input_data,
+                "output_data": output_data,
+            },
+        )
+
+
+        # Generate custom validator prompt
+        self.console.post_optimizer_status(StageType.SHOULD_OPTIMIZE)
+        validator_prompt = self.prompt_generator._generate_validator_prompt(
+            op_config, input_data, output_data
+        )
+
+        # Log the validator prompt
+        self.console.log("[bold]Validator Prompt:[/bold]")
+        self.console.log(validator_prompt)
+        self.console.log("\n")  # Add a newline for better readability
+
+        # Step 2: Use the validator prompt to assess the operation's performance
+        assessment = self.evaluator._assess_operation(
+            op_config, input_data, output_data, validator_prompt
+        )
+
+        # Print out the assessment
+        self.console.log(
+            f"[bold]Assessment for whether we should improve operation {op_config['name']}:[/bold]"
+        )
+        for key, value in assessment.items():
+            self.console.print(
+                f"[bold cyan]{key}:[/bold cyan] [yellow]{value}[/yellow]"
+            )
+        self.console.log("\n")  # Add a newline for better readability
+
+        self.runner.captured_output.save_optimizer_output(
+            stage_type=StageType.SHOULD_OPTIMIZE,
+            output={
+                "validator_prompt": validator_prompt,
+                "needs_improvement": assessment.get("needs_improvement", True),
+                "reasons": assessment.get("reasons", []),
+                "improvements": assessment.get("improvements", []),
+            },
+        )
+        self.console.post_optimizer_rationale(
+            assessment.get("needs_improvement", True),
+            "\n".join(assessment.get("reasons", []))
+            + "\n\n"
+            + "\n".join(assessment.get("improvements", [])),
+            validator_prompt,
+        )
+
+        return input_data, output_data, model_input_context_length, no_change_runtime, validator_prompt, assessment, data_exceeds_limit
+
+
+    def optimize(
+        self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]
+    ) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]:
+        """
+        Optimize the given operation configuration for the input data.
+        This method analyzes the operation and input data, generates various
+        optimization plans, evaluates them, and returns the best plan along
+        with its output. A key part of this process is creating a custom
+        validator prompt for evaluation. The validator prompt is generated
+        based on the specific task, input data, and output data. It serves
+        as a critical tool for assessing the quality and correctness of
+        each optimization plan's output. This custom prompt ensures that
+        the evaluation is tailored to the unique requirements and nuances
+        of the given operation. The types of optimization plans include:
+
+        1. Improved Prompt Plan: Enhances the original prompt based on evaluation, aiming to improve output quality.
+
+        2. Chunk Size Plan: Splits input data into chunks of different sizes,
+           processes each chunk separately, and then combines the results. This
+           can improve performance for large inputs.
+
+        3. Gleaning Plans: Implements an iterative refinement process where the
+           output is validated and improved over multiple rounds, enhancing accuracy.
+
+        4. Chain Decomposition Plan: Breaks down complex operations into a series
+           of simpler sub-operations, potentially improving overall performance
+           and interpretability.
+
+        5. Parallel Map Plan: Decomposes the task into subtasks that can be
+           executed in parallel, potentially speeding up processing for
+           independent operations.
+
+        The method generates these plans, evaluates their performance using
+        a custom validator, and selects the best performing plan based on
+        output quality and execution time.
+
+        Args:
+            op_config (Dict[str, Any]): The configuration of the operation to optimize.
+            input_data (List[Dict[str, Any]]): The input data for the operation.
+
+        Returns:
+            Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]: A tuple containing
+            the best optimization plan and its output. The plan is a list of
+            operation configurations that achieve the best performance.
+            The cost is the cost of the optimizer (from possibly synthesizing resolves).
+
+        """
+        input_data, output_data, model_input_context_length, no_change_runtime, validator_prompt, assessment, data_exceeds_limit = self._should_optimize_helper(op_config, input_data)
+
+        # Check if improvement is needed based on the assessment
+        if not self.config.get("optimizer_config", {}).get("force_decompose", False):
+            if not data_exceeds_limit and not assessment.get("needs_improvement", True):
+                self.console.log(
+                    f"[green]No improvement needed for operation {op_config['name']}[/green]"
+                )
+                return [op_config], output_data, self.plan_generator.reduce_optimizer_cost
+
+        candidate_plans = {}
+
+        # Generate improved prompt plan
+        if not data_exceeds_limit:
+            #     improved_prompt_plan = self.prompt_generator._get_improved_prompt(
+            #         op_config, assessment, input_data
+            #     )
+            #     candidate_plans["improved_instructions"] = improved_prompt_plan
+            candidate_plans["no_change"] = [op_config]
+
+        # Generate chunk size plans
+        self.console.post_optimizer_status(StageType.CANDIDATE_PLANS)
+        self.console.log("[bold magenta]Generating chunking plans...[/bold magenta]")
+        chunk_size_plans = self.plan_generator._generate_chunk_size_plans(
+            op_config, input_data, validator_prompt, model_input_context_length
+        )
+        for pname, plan in chunk_size_plans.items():
+            candidate_plans[pname] = plan
+
+        # Generate gleaning plans
+        if not data_exceeds_limit:
+            self.console.log(
+                "[bold magenta]Generating gleaning plans...[/bold magenta]"
+            )
+            gleaning_plans = self.plan_generator._generate_gleaning_plans(
+                op_config, validator_prompt
+            )
+            for pname, plan in gleaning_plans.items():
+                candidate_plans[pname] = plan
+
+        # Generate chain decomposition plans
+        if not data_exceeds_limit:
+            if not self.is_filter:
+                self.console.log(
+                    "[bold magenta]Generating chain projection synthesis plans...[/bold magenta]"
+                )
+                chain_plans = self.plan_generator._generate_chain_plans(
+                    op_config, input_data
+                )
+                for pname, plan in chain_plans.items():
+                    candidate_plans[pname] = plan
+
+                # Generate parallel map plans
+                self.console.log(
+                    "[bold magenta]Generating independent projection synthesis plans...[/bold magenta]"
+                )
+                parallel_plans = self.plan_generator._generate_parallel_plans(
+                    op_config, input_data
+                )
+                for pname, plan in parallel_plans.items():
+                    candidate_plans[pname] = plan
+
+        # Select consistent evaluation samples
+        num_evaluations = min(5, len(input_data))
+        evaluation_samples = select_evaluation_samples(input_data, num_evaluations)
+
+        results = {}
+        plans_list = list(candidate_plans.items())
+
+        # Capture candidate plans
+        self.runner.captured_output.save_optimizer_output(
+            stage_type=StageType.CANDIDATE_PLANS,
+            output=candidate_plans,
+        )
+
+        self.console.post_optimizer_status(StageType.EVALUATION_RESULTS)
+        self.console.log(
+            f"[bold magenta]Evaluating {len(plans_list)} plans...[/bold magenta]"
+        )
+        for i in range(0, len(plans_list), self._num_plans_to_evaluate_in_parallel):
+            batch = plans_list[i : i + self._num_plans_to_evaluate_in_parallel]
+            with ThreadPoolExecutor(
+                max_workers=self._num_plans_to_evaluate_in_parallel
+            ) as executor:
+                futures = {
+                    executor.submit(
+                        self.evaluator._evaluate_plan,
+                        plan_name,
+                        op_config,
+                        plan,
+                        copy.deepcopy(evaluation_samples),
+                        validator_prompt,
+                    ): plan_name
+                    for plan_name, plan in batch
+                }
+                for future in as_completed(futures):
+                    plan_name = futures[future]
+                    try:
+                        score, runtime, output = future.result(timeout=self.timeout)
+                        results[plan_name] = (score, runtime, output)
+                    except concurrent.futures.TimeoutError:
+                        self.console.log(
+                            f"[yellow]Plan {plan_name} timed out and will be skipped.[/yellow]"
+                        )
+                    except Exception as e:
+                        # TODO: raise this error if the error is related to a Jinja error
+                        self.console.log(
+                            f"[red]Error in plan {plan_name}: {str(e)}[/red]"
+                        )
+                        import traceback
+
+                        print(traceback.format_exc())
+
+        # Add no change plan
+        if not data_exceeds_limit:
+            results["no_change"] = (
+                results["no_change"][0],
+                no_change_runtime,
+                results["no_change"][2],
+            )
+
+        # Create a table of scores sorted in descending order
+        scores = sorted(
+            [(score, runtime, plan) for plan, (score, runtime, _) in results.items()],
+            reverse=True,
+        )
+
+        # Sort results by score in descending order
+        sorted_results = sorted(results.items(), key=lambda x: x[1][0], reverse=True)
+
+        # Take the top 6 plans
+        top_plans = sorted_results[: self.k_to_pairwise_compare]
+
+        # Check if there are no top plans
+        if len(top_plans) == 0:
+            self.console.post_optimizer_status(StageType.END)
+            raise ValueError(
+                "Agent did not generate any plans. Unable to proceed with optimization. Try again."
+            )
+
+        # Include any additional plans that are tied with the last plan
+        tail_score = (
+            top_plans[-1][1][0]
+            if len(top_plans) == self.k_to_pairwise_compare
+            else float("-inf")
+        )
+        filtered_results = dict(
+            top_plans
+            + [
+                item
+                for item in sorted_results[len(top_plans) :]
+                if item[1][0] == tail_score
+            ]
+        )
+
+        # Perform pairwise comparisons on filtered plans
+        if len(filtered_results) > 1:
+            pairwise_rankings = self.evaluator._pairwise_compare_plans(
+                filtered_results, validator_prompt, op_config, evaluation_samples
+            )
+            best_plan_name = max(pairwise_rankings, key=pairwise_rankings.get)
+        else:
+            pairwise_rankings = {k: 0 for k in results.keys()}
+            best_plan_name = (
+                next(iter(filtered_results))
+                if filtered_results
+                else max(results, key=lambda x: results[x][0])
+            )
+
+        self.console.log(
+            f"\n[bold]Plan Evaluation Results for {op_config['name']} ({op_config['type']}, {len(scores)} plans, {num_evaluations} samples):[/bold]"
+        )
+        table = Table(show_header=True, header_style="bold magenta")
+        table.add_column("Plan", style="dim")
+        table.add_column("Score", justify="right", width=10)
+        table.add_column("Runtime", justify="right", width=10)
+        table.add_column("Pairwise Wins", justify="right", width=10)
+
+        for score, runtime, plan in scores:
+            table.add_row(
+                plan,
+                f"{score:.2f}",
+                f"{runtime:.2f}s",
+                f"{pairwise_rankings.get(plan, 0)}",
+            )
+
+        self.console.log(table)
+        self.console.log("\n")
+
+        _, _, best_output = results[best_plan_name]
+        self.console.log(
+            f"[green]Choosing {best_plan_name} for operation {op_config['name']} (Score: {results[best_plan_name][0]:.2f}, Runtime: {results[best_plan_name][1]:.2f}s)[/green]"
+        )
+
+        # Capture evaluation results
+        ratings = {k: v[0] for k, v in results.items()}
+        runtime = {k: v[1] for k, v in results.items()}
+        sample_outputs = {k: v[2] for k, v in results.items()}
+        self.runner.captured_output.save_optimizer_output(
+            stage_type=StageType.EVALUATION_RESULTS,
+            output={
+                "input_data": evaluation_samples,
+                "all_plan_ratings": ratings,
+                "all_plan_runtimes": runtime,
+                "all_plan_sample_outputs": sample_outputs,
+                "all_plan_pairwise_rankings": pairwise_rankings,
+            },
+        )
+
+        self.console.post_optimizer_status(StageType.END)
+        return (
+            candidate_plans[best_plan_name],
+            best_output,
+            self.plan_generator.reduce_optimizer_cost,
+        )
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ __init__(runner, config, console, llm_client, max_threads, run_operation, timeout=10, is_filter=False) + +

+ + +
+ +

Initialize the MapOptimizer.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ config + + Dict[str, Any] + +
+

The configuration dictionary for the optimizer.

+
+
+ required +
+ console + + Console + +
+

A Rich console object for pretty printing.

+
+
+ required +
+ llm_client + + LLMClient + +
+

A client for interacting with a language model.

+
+
+ required +
+ max_threads + + int + +
+

The maximum number of threads to use for parallel execution.

+
+
+ required +
+ run_operation + + Callable + +
+

A function to execute operations.

+
+
+ required +
+ timeout + + int + +
+

The timeout in seconds for operation execution. Defaults to 10.

+
+
+ 10 +
+ is_filter + + bool + +
+

If True, the operation is a filter operation. Defaults to False.

+
+
+ False +
+ +
+ Source code in docetl/optimizers/map_optimizer/optimizer.py +
40
+41
+42
+43
+44
+45
+46
+47
+48
+49
+50
+51
+52
+53
+54
+55
+56
+57
+58
+59
+60
+61
+62
+63
+64
+65
+66
+67
+68
+69
+70
+71
+72
+73
+74
+75
+76
+77
+78
+79
+80
+81
+82
+83
+84
+85
+86
+87
def __init__(
+    self,
+    runner,
+    config: Dict[str, Any],
+    console: Console,
+    llm_client: LLMClient,
+    max_threads: int,
+    run_operation: Callable,
+    timeout: int = 10,
+    is_filter: bool = False,
+):
+    """
+    Initialize the MapOptimizer.
+
+    Args:
+        config (Dict[str, Any]): The configuration dictionary for the optimizer.
+        console (Console): A Rich console object for pretty printing.
+        llm_client (LLMClient): A client for interacting with a language model.
+        max_threads (int): The maximum number of threads to use for parallel execution.
+        run_operation (Callable): A function to execute operations.
+        timeout (int, optional): The timeout in seconds for operation execution. Defaults to 10.
+        is_filter (bool, optional): If True, the operation is a filter operation. Defaults to False.
+    """
+    self.runner = runner
+    self.config = config
+    self.console = console
+    self.llm_client = llm_client
+    self._run_operation = run_operation
+    self.max_threads = max_threads
+    self.timeout = timeout
+    self._num_plans_to_evaluate_in_parallel = 5
+    self.is_filter = is_filter
+    self.k_to_pairwise_compare = 6
+
+    self.plan_generator = PlanGenerator(
+        runner, llm_client, console, config, run_operation, max_threads, is_filter
+    )
+    self.evaluator = Evaluator(
+        llm_client,
+        console,
+        run_operation,
+        timeout,
+        self._num_plans_to_evaluate_in_parallel,
+        is_filter,
+    )
+    self.prompt_generator = PromptGenerator(
+        runner, llm_client, console, config, max_threads, is_filter
+    )
+
+
+
+ +
+ +
+ + +

+ optimize(op_config, input_data) + +

+ + +
+ +

Optimize the given operation configuration for the input data. +This method analyzes the operation and input data, generates various +optimization plans, evaluates them, and returns the best plan along +with its output. A key part of this process is creating a custom +validator prompt for evaluation. The validator prompt is generated +based on the specific task, input data, and output data. It serves +as a critical tool for assessing the quality and correctness of +each optimization plan's output. This custom prompt ensures that +the evaluation is tailored to the unique requirements and nuances +of the given operation. The types of optimization plans include:

+
    +
  1. +

    Improved Prompt Plan: Enhances the original prompt based on evaluation, aiming to improve output quality.

    +
  2. +
  3. +

    Chunk Size Plan: Splits input data into chunks of different sizes, + processes each chunk separately, and then combines the results. This + can improve performance for large inputs.

    +
  4. +
  5. +

    Gleaning Plans: Implements an iterative refinement process where the + output is validated and improved over multiple rounds, enhancing accuracy.

    +
  6. +
  7. +

    Chain Decomposition Plan: Breaks down complex operations into a series + of simpler sub-operations, potentially improving overall performance + and interpretability.

    +
  8. +
  9. +

    Parallel Map Plan: Decomposes the task into subtasks that can be + executed in parallel, potentially speeding up processing for + independent operations.

    +
  10. +
+

The method generates these plans, evaluates their performance using +a custom validator, and selects the best performing plan based on +output quality and execution time.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ op_config + + Dict[str, Any] + +
+

The configuration of the operation to optimize.

+
+
+ required +
+ input_data + + List[Dict[str, Any]] + +
+

The input data for the operation.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + + + + + + + + + + + + + +
TypeDescription
+ List[Dict[str, Any]] + +
+

Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]: A tuple containing

+
+
+ List[Dict[str, Any]] + +
+

the best optimization plan and its output. The plan is a list of

+
+
+ float + +
+

operation configurations that achieve the best performance.

+
+
+ Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float] + +
+

The cost is the cost of the optimizer (from possibly synthesizing resolves).

+
+
+ +
+ Source code in docetl/optimizers/map_optimizer/optimizer.py +
208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
+338
+339
+340
+341
+342
+343
+344
+345
+346
+347
+348
+349
+350
+351
+352
+353
+354
+355
+356
+357
+358
+359
+360
+361
+362
+363
+364
+365
+366
+367
+368
+369
+370
+371
+372
+373
+374
+375
+376
+377
+378
+379
+380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
+438
+439
+440
+441
+442
+443
+444
+445
+446
+447
+448
+449
+450
+451
+452
+453
+454
+455
+456
+457
+458
+459
+460
+461
+462
+463
+464
+465
+466
+467
+468
+469
def optimize(
+    self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]
+) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]:
+    """
+    Optimize the given operation configuration for the input data.
+    This method analyzes the operation and input data, generates various
+    optimization plans, evaluates them, and returns the best plan along
+    with its output. A key part of this process is creating a custom
+    validator prompt for evaluation. The validator prompt is generated
+    based on the specific task, input data, and output data. It serves
+    as a critical tool for assessing the quality and correctness of
+    each optimization plan's output. This custom prompt ensures that
+    the evaluation is tailored to the unique requirements and nuances
+    of the given operation. The types of optimization plans include:
+
+    1. Improved Prompt Plan: Enhances the original prompt based on evaluation, aiming to improve output quality.
+
+    2. Chunk Size Plan: Splits input data into chunks of different sizes,
+       processes each chunk separately, and then combines the results. This
+       can improve performance for large inputs.
+
+    3. Gleaning Plans: Implements an iterative refinement process where the
+       output is validated and improved over multiple rounds, enhancing accuracy.
+
+    4. Chain Decomposition Plan: Breaks down complex operations into a series
+       of simpler sub-operations, potentially improving overall performance
+       and interpretability.
+
+    5. Parallel Map Plan: Decomposes the task into subtasks that can be
+       executed in parallel, potentially speeding up processing for
+       independent operations.
+
+    The method generates these plans, evaluates their performance using
+    a custom validator, and selects the best performing plan based on
+    output quality and execution time.
+
+    Args:
+        op_config (Dict[str, Any]): The configuration of the operation to optimize.
+        input_data (List[Dict[str, Any]]): The input data for the operation.
+
+    Returns:
+        Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]: A tuple containing
+        the best optimization plan and its output. The plan is a list of
+        operation configurations that achieve the best performance.
+        The cost is the cost of the optimizer (from possibly synthesizing resolves).
+
+    """
+    input_data, output_data, model_input_context_length, no_change_runtime, validator_prompt, assessment, data_exceeds_limit = self._should_optimize_helper(op_config, input_data)
+
+    # Check if improvement is needed based on the assessment
+    if not self.config.get("optimizer_config", {}).get("force_decompose", False):
+        if not data_exceeds_limit and not assessment.get("needs_improvement", True):
+            self.console.log(
+                f"[green]No improvement needed for operation {op_config['name']}[/green]"
+            )
+            return [op_config], output_data, self.plan_generator.reduce_optimizer_cost
+
+    candidate_plans = {}
+
+    # Generate improved prompt plan
+    if not data_exceeds_limit:
+        #     improved_prompt_plan = self.prompt_generator._get_improved_prompt(
+        #         op_config, assessment, input_data
+        #     )
+        #     candidate_plans["improved_instructions"] = improved_prompt_plan
+        candidate_plans["no_change"] = [op_config]
+
+    # Generate chunk size plans
+    self.console.post_optimizer_status(StageType.CANDIDATE_PLANS)
+    self.console.log("[bold magenta]Generating chunking plans...[/bold magenta]")
+    chunk_size_plans = self.plan_generator._generate_chunk_size_plans(
+        op_config, input_data, validator_prompt, model_input_context_length
+    )
+    for pname, plan in chunk_size_plans.items():
+        candidate_plans[pname] = plan
+
+    # Generate gleaning plans
+    if not data_exceeds_limit:
+        self.console.log(
+            "[bold magenta]Generating gleaning plans...[/bold magenta]"
+        )
+        gleaning_plans = self.plan_generator._generate_gleaning_plans(
+            op_config, validator_prompt
+        )
+        for pname, plan in gleaning_plans.items():
+            candidate_plans[pname] = plan
+
+    # Generate chain decomposition plans
+    if not data_exceeds_limit:
+        if not self.is_filter:
+            self.console.log(
+                "[bold magenta]Generating chain projection synthesis plans...[/bold magenta]"
+            )
+            chain_plans = self.plan_generator._generate_chain_plans(
+                op_config, input_data
+            )
+            for pname, plan in chain_plans.items():
+                candidate_plans[pname] = plan
+
+            # Generate parallel map plans
+            self.console.log(
+                "[bold magenta]Generating independent projection synthesis plans...[/bold magenta]"
+            )
+            parallel_plans = self.plan_generator._generate_parallel_plans(
+                op_config, input_data
+            )
+            for pname, plan in parallel_plans.items():
+                candidate_plans[pname] = plan
+
+    # Select consistent evaluation samples
+    num_evaluations = min(5, len(input_data))
+    evaluation_samples = select_evaluation_samples(input_data, num_evaluations)
+
+    results = {}
+    plans_list = list(candidate_plans.items())
+
+    # Capture candidate plans
+    self.runner.captured_output.save_optimizer_output(
+        stage_type=StageType.CANDIDATE_PLANS,
+        output=candidate_plans,
+    )
+
+    self.console.post_optimizer_status(StageType.EVALUATION_RESULTS)
+    self.console.log(
+        f"[bold magenta]Evaluating {len(plans_list)} plans...[/bold magenta]"
+    )
+    for i in range(0, len(plans_list), self._num_plans_to_evaluate_in_parallel):
+        batch = plans_list[i : i + self._num_plans_to_evaluate_in_parallel]
+        with ThreadPoolExecutor(
+            max_workers=self._num_plans_to_evaluate_in_parallel
+        ) as executor:
+            futures = {
+                executor.submit(
+                    self.evaluator._evaluate_plan,
+                    plan_name,
+                    op_config,
+                    plan,
+                    copy.deepcopy(evaluation_samples),
+                    validator_prompt,
+                ): plan_name
+                for plan_name, plan in batch
+            }
+            for future in as_completed(futures):
+                plan_name = futures[future]
+                try:
+                    score, runtime, output = future.result(timeout=self.timeout)
+                    results[plan_name] = (score, runtime, output)
+                except concurrent.futures.TimeoutError:
+                    self.console.log(
+                        f"[yellow]Plan {plan_name} timed out and will be skipped.[/yellow]"
+                    )
+                except Exception as e:
+                    # TODO: raise this error if the error is related to a Jinja error
+                    self.console.log(
+                        f"[red]Error in plan {plan_name}: {str(e)}[/red]"
+                    )
+                    import traceback
+
+                    print(traceback.format_exc())
+
+    # Add no change plan
+    if not data_exceeds_limit:
+        results["no_change"] = (
+            results["no_change"][0],
+            no_change_runtime,
+            results["no_change"][2],
+        )
+
+    # Create a table of scores sorted in descending order
+    scores = sorted(
+        [(score, runtime, plan) for plan, (score, runtime, _) in results.items()],
+        reverse=True,
+    )
+
+    # Sort results by score in descending order
+    sorted_results = sorted(results.items(), key=lambda x: x[1][0], reverse=True)
+
+    # Take the top 6 plans
+    top_plans = sorted_results[: self.k_to_pairwise_compare]
+
+    # Check if there are no top plans
+    if len(top_plans) == 0:
+        self.console.post_optimizer_status(StageType.END)
+        raise ValueError(
+            "Agent did not generate any plans. Unable to proceed with optimization. Try again."
+        )
+
+    # Include any additional plans that are tied with the last plan
+    tail_score = (
+        top_plans[-1][1][0]
+        if len(top_plans) == self.k_to_pairwise_compare
+        else float("-inf")
+    )
+    filtered_results = dict(
+        top_plans
+        + [
+            item
+            for item in sorted_results[len(top_plans) :]
+            if item[1][0] == tail_score
+        ]
+    )
+
+    # Perform pairwise comparisons on filtered plans
+    if len(filtered_results) > 1:
+        pairwise_rankings = self.evaluator._pairwise_compare_plans(
+            filtered_results, validator_prompt, op_config, evaluation_samples
+        )
+        best_plan_name = max(pairwise_rankings, key=pairwise_rankings.get)
+    else:
+        pairwise_rankings = {k: 0 for k in results.keys()}
+        best_plan_name = (
+            next(iter(filtered_results))
+            if filtered_results
+            else max(results, key=lambda x: results[x][0])
+        )
+
+    self.console.log(
+        f"\n[bold]Plan Evaluation Results for {op_config['name']} ({op_config['type']}, {len(scores)} plans, {num_evaluations} samples):[/bold]"
+    )
+    table = Table(show_header=True, header_style="bold magenta")
+    table.add_column("Plan", style="dim")
+    table.add_column("Score", justify="right", width=10)
+    table.add_column("Runtime", justify="right", width=10)
+    table.add_column("Pairwise Wins", justify="right", width=10)
+
+    for score, runtime, plan in scores:
+        table.add_row(
+            plan,
+            f"{score:.2f}",
+            f"{runtime:.2f}s",
+            f"{pairwise_rankings.get(plan, 0)}",
+        )
+
+    self.console.log(table)
+    self.console.log("\n")
+
+    _, _, best_output = results[best_plan_name]
+    self.console.log(
+        f"[green]Choosing {best_plan_name} for operation {op_config['name']} (Score: {results[best_plan_name][0]:.2f}, Runtime: {results[best_plan_name][1]:.2f}s)[/green]"
+    )
+
+    # Capture evaluation results
+    ratings = {k: v[0] for k, v in results.items()}
+    runtime = {k: v[1] for k, v in results.items()}
+    sample_outputs = {k: v[2] for k, v in results.items()}
+    self.runner.captured_output.save_optimizer_output(
+        stage_type=StageType.EVALUATION_RESULTS,
+        output={
+            "input_data": evaluation_samples,
+            "all_plan_ratings": ratings,
+            "all_plan_runtimes": runtime,
+            "all_plan_sample_outputs": sample_outputs,
+            "all_plan_pairwise_rankings": pairwise_rankings,
+        },
+    )
+
+    self.console.post_optimizer_status(StageType.END)
+    return (
+        candidate_plans[best_plan_name],
+        best_output,
+        self.plan_generator.reduce_optimizer_cost,
+    )
+
+
+
+ +
+ +
+ + +

+ should_optimize(op_config, input_data) + +

+ + +
+ +

Determine if the given operation configuration should be optimized.

+ +
+ Source code in docetl/optimizers/map_optimizer/optimizer.py +
 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
def should_optimize(self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]) -> Tuple[str, List[Dict[str, Any]], List[Dict[str, Any]]]:
+    """
+    Determine if the given operation configuration should be optimized.
+    """
+    input_data, output_data, _, _, validator_prompt, assessment, data_exceeds_limit = self._should_optimize_helper(op_config, input_data)
+    if data_exceeds_limit or assessment.get("needs_improvement", True):
+        assessment_str = "\n".join(assessment.get("reasons", [])) + "\n\nHere are some improvements that may help:\n" + "\n".join(assessment.get("improvements", []))
+        if data_exceeds_limit:
+            assessment_str += "\nAlso, the input data exceeds the token limit."
+        return assessment_str, input_data, output_data
+    else:
+        return "", input_data, output_data
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.optimizers.reduce_optimizer.ReduceOptimizer + + +

+ + +
+ + +

A class that optimizes reduce operations in data processing pipelines.

+

This optimizer analyzes the input and output of a reduce operation, creates and evaluates +multiple reduce plans, and selects the best plan for optimizing the operation's performance.

+ + +

Attributes:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescription
config + Dict[str, Any] + +
+

Configuration dictionary for the optimizer.

+
+
console + Console + +
+

Rich console object for pretty printing.

+
+
llm_client + LLMClient + +
+

Client for interacting with a language model.

+
+
_run_operation + Callable + +
+

Function to run an operation.

+
+
max_threads + int + +
+

Maximum number of threads to use for parallel processing.

+
+
num_fold_prompts + int + +
+

Number of fold prompts to generate.

+
+
num_samples_in_validation + int + +
+

Number of samples to use in validation.

+
+
+ + + + + + +
+ Source code in docetl/optimizers/reduce_optimizer.py +
  21
+  22
+  23
+  24
+  25
+  26
+  27
+  28
+  29
+  30
+  31
+  32
+  33
+  34
+  35
+  36
+  37
+  38
+  39
+  40
+  41
+  42
+  43
+  44
+  45
+  46
+  47
+  48
+  49
+  50
+  51
+  52
+  53
+  54
+  55
+  56
+  57
+  58
+  59
+  60
+  61
+  62
+  63
+  64
+  65
+  66
+  67
+  68
+  69
+  70
+  71
+  72
+  73
+  74
+  75
+  76
+  77
+  78
+  79
+  80
+  81
+  82
+  83
+  84
+  85
+  86
+  87
+  88
+  89
+  90
+  91
+  92
+  93
+  94
+  95
+  96
+  97
+  98
+  99
+ 100
+ 101
+ 102
+ 103
+ 104
+ 105
+ 106
+ 107
+ 108
+ 109
+ 110
+ 111
+ 112
+ 113
+ 114
+ 115
+ 116
+ 117
+ 118
+ 119
+ 120
+ 121
+ 122
+ 123
+ 124
+ 125
+ 126
+ 127
+ 128
+ 129
+ 130
+ 131
+ 132
+ 133
+ 134
+ 135
+ 136
+ 137
+ 138
+ 139
+ 140
+ 141
+ 142
+ 143
+ 144
+ 145
+ 146
+ 147
+ 148
+ 149
+ 150
+ 151
+ 152
+ 153
+ 154
+ 155
+ 156
+ 157
+ 158
+ 159
+ 160
+ 161
+ 162
+ 163
+ 164
+ 165
+ 166
+ 167
+ 168
+ 169
+ 170
+ 171
+ 172
+ 173
+ 174
+ 175
+ 176
+ 177
+ 178
+ 179
+ 180
+ 181
+ 182
+ 183
+ 184
+ 185
+ 186
+ 187
+ 188
+ 189
+ 190
+ 191
+ 192
+ 193
+ 194
+ 195
+ 196
+ 197
+ 198
+ 199
+ 200
+ 201
+ 202
+ 203
+ 204
+ 205
+ 206
+ 207
+ 208
+ 209
+ 210
+ 211
+ 212
+ 213
+ 214
+ 215
+ 216
+ 217
+ 218
+ 219
+ 220
+ 221
+ 222
+ 223
+ 224
+ 225
+ 226
+ 227
+ 228
+ 229
+ 230
+ 231
+ 232
+ 233
+ 234
+ 235
+ 236
+ 237
+ 238
+ 239
+ 240
+ 241
+ 242
+ 243
+ 244
+ 245
+ 246
+ 247
+ 248
+ 249
+ 250
+ 251
+ 252
+ 253
+ 254
+ 255
+ 256
+ 257
+ 258
+ 259
+ 260
+ 261
+ 262
+ 263
+ 264
+ 265
+ 266
+ 267
+ 268
+ 269
+ 270
+ 271
+ 272
+ 273
+ 274
+ 275
+ 276
+ 277
+ 278
+ 279
+ 280
+ 281
+ 282
+ 283
+ 284
+ 285
+ 286
+ 287
+ 288
+ 289
+ 290
+ 291
+ 292
+ 293
+ 294
+ 295
+ 296
+ 297
+ 298
+ 299
+ 300
+ 301
+ 302
+ 303
+ 304
+ 305
+ 306
+ 307
+ 308
+ 309
+ 310
+ 311
+ 312
+ 313
+ 314
+ 315
+ 316
+ 317
+ 318
+ 319
+ 320
+ 321
+ 322
+ 323
+ 324
+ 325
+ 326
+ 327
+ 328
+ 329
+ 330
+ 331
+ 332
+ 333
+ 334
+ 335
+ 336
+ 337
+ 338
+ 339
+ 340
+ 341
+ 342
+ 343
+ 344
+ 345
+ 346
+ 347
+ 348
+ 349
+ 350
+ 351
+ 352
+ 353
+ 354
+ 355
+ 356
+ 357
+ 358
+ 359
+ 360
+ 361
+ 362
+ 363
+ 364
+ 365
+ 366
+ 367
+ 368
+ 369
+ 370
+ 371
+ 372
+ 373
+ 374
+ 375
+ 376
+ 377
+ 378
+ 379
+ 380
+ 381
+ 382
+ 383
+ 384
+ 385
+ 386
+ 387
+ 388
+ 389
+ 390
+ 391
+ 392
+ 393
+ 394
+ 395
+ 396
+ 397
+ 398
+ 399
+ 400
+ 401
+ 402
+ 403
+ 404
+ 405
+ 406
+ 407
+ 408
+ 409
+ 410
+ 411
+ 412
+ 413
+ 414
+ 415
+ 416
+ 417
+ 418
+ 419
+ 420
+ 421
+ 422
+ 423
+ 424
+ 425
+ 426
+ 427
+ 428
+ 429
+ 430
+ 431
+ 432
+ 433
+ 434
+ 435
+ 436
+ 437
+ 438
+ 439
+ 440
+ 441
+ 442
+ 443
+ 444
+ 445
+ 446
+ 447
+ 448
+ 449
+ 450
+ 451
+ 452
+ 453
+ 454
+ 455
+ 456
+ 457
+ 458
+ 459
+ 460
+ 461
+ 462
+ 463
+ 464
+ 465
+ 466
+ 467
+ 468
+ 469
+ 470
+ 471
+ 472
+ 473
+ 474
+ 475
+ 476
+ 477
+ 478
+ 479
+ 480
+ 481
+ 482
+ 483
+ 484
+ 485
+ 486
+ 487
+ 488
+ 489
+ 490
+ 491
+ 492
+ 493
+ 494
+ 495
+ 496
+ 497
+ 498
+ 499
+ 500
+ 501
+ 502
+ 503
+ 504
+ 505
+ 506
+ 507
+ 508
+ 509
+ 510
+ 511
+ 512
+ 513
+ 514
+ 515
+ 516
+ 517
+ 518
+ 519
+ 520
+ 521
+ 522
+ 523
+ 524
+ 525
+ 526
+ 527
+ 528
+ 529
+ 530
+ 531
+ 532
+ 533
+ 534
+ 535
+ 536
+ 537
+ 538
+ 539
+ 540
+ 541
+ 542
+ 543
+ 544
+ 545
+ 546
+ 547
+ 548
+ 549
+ 550
+ 551
+ 552
+ 553
+ 554
+ 555
+ 556
+ 557
+ 558
+ 559
+ 560
+ 561
+ 562
+ 563
+ 564
+ 565
+ 566
+ 567
+ 568
+ 569
+ 570
+ 571
+ 572
+ 573
+ 574
+ 575
+ 576
+ 577
+ 578
+ 579
+ 580
+ 581
+ 582
+ 583
+ 584
+ 585
+ 586
+ 587
+ 588
+ 589
+ 590
+ 591
+ 592
+ 593
+ 594
+ 595
+ 596
+ 597
+ 598
+ 599
+ 600
+ 601
+ 602
+ 603
+ 604
+ 605
+ 606
+ 607
+ 608
+ 609
+ 610
+ 611
+ 612
+ 613
+ 614
+ 615
+ 616
+ 617
+ 618
+ 619
+ 620
+ 621
+ 622
+ 623
+ 624
+ 625
+ 626
+ 627
+ 628
+ 629
+ 630
+ 631
+ 632
+ 633
+ 634
+ 635
+ 636
+ 637
+ 638
+ 639
+ 640
+ 641
+ 642
+ 643
+ 644
+ 645
+ 646
+ 647
+ 648
+ 649
+ 650
+ 651
+ 652
+ 653
+ 654
+ 655
+ 656
+ 657
+ 658
+ 659
+ 660
+ 661
+ 662
+ 663
+ 664
+ 665
+ 666
+ 667
+ 668
+ 669
+ 670
+ 671
+ 672
+ 673
+ 674
+ 675
+ 676
+ 677
+ 678
+ 679
+ 680
+ 681
+ 682
+ 683
+ 684
+ 685
+ 686
+ 687
+ 688
+ 689
+ 690
+ 691
+ 692
+ 693
+ 694
+ 695
+ 696
+ 697
+ 698
+ 699
+ 700
+ 701
+ 702
+ 703
+ 704
+ 705
+ 706
+ 707
+ 708
+ 709
+ 710
+ 711
+ 712
+ 713
+ 714
+ 715
+ 716
+ 717
+ 718
+ 719
+ 720
+ 721
+ 722
+ 723
+ 724
+ 725
+ 726
+ 727
+ 728
+ 729
+ 730
+ 731
+ 732
+ 733
+ 734
+ 735
+ 736
+ 737
+ 738
+ 739
+ 740
+ 741
+ 742
+ 743
+ 744
+ 745
+ 746
+ 747
+ 748
+ 749
+ 750
+ 751
+ 752
+ 753
+ 754
+ 755
+ 756
+ 757
+ 758
+ 759
+ 760
+ 761
+ 762
+ 763
+ 764
+ 765
+ 766
+ 767
+ 768
+ 769
+ 770
+ 771
+ 772
+ 773
+ 774
+ 775
+ 776
+ 777
+ 778
+ 779
+ 780
+ 781
+ 782
+ 783
+ 784
+ 785
+ 786
+ 787
+ 788
+ 789
+ 790
+ 791
+ 792
+ 793
+ 794
+ 795
+ 796
+ 797
+ 798
+ 799
+ 800
+ 801
+ 802
+ 803
+ 804
+ 805
+ 806
+ 807
+ 808
+ 809
+ 810
+ 811
+ 812
+ 813
+ 814
+ 815
+ 816
+ 817
+ 818
+ 819
+ 820
+ 821
+ 822
+ 823
+ 824
+ 825
+ 826
+ 827
+ 828
+ 829
+ 830
+ 831
+ 832
+ 833
+ 834
+ 835
+ 836
+ 837
+ 838
+ 839
+ 840
+ 841
+ 842
+ 843
+ 844
+ 845
+ 846
+ 847
+ 848
+ 849
+ 850
+ 851
+ 852
+ 853
+ 854
+ 855
+ 856
+ 857
+ 858
+ 859
+ 860
+ 861
+ 862
+ 863
+ 864
+ 865
+ 866
+ 867
+ 868
+ 869
+ 870
+ 871
+ 872
+ 873
+ 874
+ 875
+ 876
+ 877
+ 878
+ 879
+ 880
+ 881
+ 882
+ 883
+ 884
+ 885
+ 886
+ 887
+ 888
+ 889
+ 890
+ 891
+ 892
+ 893
+ 894
+ 895
+ 896
+ 897
+ 898
+ 899
+ 900
+ 901
+ 902
+ 903
+ 904
+ 905
+ 906
+ 907
+ 908
+ 909
+ 910
+ 911
+ 912
+ 913
+ 914
+ 915
+ 916
+ 917
+ 918
+ 919
+ 920
+ 921
+ 922
+ 923
+ 924
+ 925
+ 926
+ 927
+ 928
+ 929
+ 930
+ 931
+ 932
+ 933
+ 934
+ 935
+ 936
+ 937
+ 938
+ 939
+ 940
+ 941
+ 942
+ 943
+ 944
+ 945
+ 946
+ 947
+ 948
+ 949
+ 950
+ 951
+ 952
+ 953
+ 954
+ 955
+ 956
+ 957
+ 958
+ 959
+ 960
+ 961
+ 962
+ 963
+ 964
+ 965
+ 966
+ 967
+ 968
+ 969
+ 970
+ 971
+ 972
+ 973
+ 974
+ 975
+ 976
+ 977
+ 978
+ 979
+ 980
+ 981
+ 982
+ 983
+ 984
+ 985
+ 986
+ 987
+ 988
+ 989
+ 990
+ 991
+ 992
+ 993
+ 994
+ 995
+ 996
+ 997
+ 998
+ 999
+1000
+1001
+1002
+1003
+1004
+1005
+1006
+1007
+1008
+1009
+1010
+1011
+1012
+1013
+1014
+1015
+1016
+1017
+1018
+1019
+1020
+1021
+1022
+1023
+1024
+1025
+1026
+1027
+1028
+1029
+1030
+1031
+1032
+1033
+1034
+1035
+1036
+1037
+1038
+1039
+1040
+1041
+1042
+1043
+1044
+1045
+1046
+1047
+1048
+1049
+1050
+1051
+1052
+1053
+1054
+1055
+1056
+1057
+1058
+1059
+1060
+1061
+1062
+1063
+1064
+1065
+1066
+1067
+1068
+1069
+1070
+1071
+1072
+1073
+1074
+1075
+1076
+1077
+1078
+1079
+1080
+1081
+1082
+1083
+1084
+1085
+1086
+1087
+1088
+1089
+1090
+1091
+1092
+1093
+1094
+1095
+1096
+1097
+1098
+1099
+1100
+1101
+1102
+1103
+1104
+1105
+1106
+1107
+1108
+1109
+1110
+1111
+1112
+1113
+1114
+1115
+1116
+1117
+1118
+1119
+1120
+1121
+1122
+1123
+1124
+1125
+1126
+1127
+1128
+1129
+1130
+1131
+1132
+1133
+1134
+1135
+1136
+1137
+1138
+1139
+1140
+1141
+1142
+1143
+1144
+1145
+1146
+1147
+1148
+1149
+1150
+1151
+1152
+1153
+1154
+1155
+1156
+1157
+1158
+1159
+1160
+1161
+1162
+1163
+1164
+1165
+1166
+1167
+1168
+1169
+1170
+1171
+1172
+1173
+1174
+1175
+1176
+1177
+1178
+1179
+1180
+1181
+1182
+1183
+1184
+1185
+1186
+1187
+1188
+1189
+1190
+1191
+1192
+1193
+1194
+1195
+1196
+1197
+1198
+1199
+1200
+1201
+1202
+1203
+1204
+1205
+1206
+1207
+1208
+1209
+1210
+1211
+1212
+1213
+1214
+1215
+1216
+1217
+1218
+1219
+1220
+1221
+1222
+1223
+1224
+1225
+1226
+1227
+1228
+1229
+1230
+1231
+1232
+1233
+1234
+1235
+1236
+1237
+1238
+1239
+1240
+1241
+1242
+1243
+1244
+1245
+1246
+1247
+1248
+1249
+1250
+1251
+1252
+1253
+1254
+1255
+1256
+1257
+1258
+1259
+1260
+1261
+1262
+1263
+1264
+1265
+1266
+1267
+1268
+1269
+1270
+1271
+1272
+1273
+1274
+1275
+1276
+1277
+1278
+1279
+1280
+1281
+1282
+1283
+1284
+1285
+1286
+1287
+1288
+1289
+1290
+1291
+1292
+1293
+1294
+1295
+1296
+1297
+1298
+1299
+1300
+1301
+1302
+1303
+1304
+1305
+1306
+1307
+1308
+1309
+1310
+1311
+1312
+1313
+1314
+1315
+1316
+1317
+1318
+1319
+1320
+1321
+1322
+1323
+1324
+1325
+1326
+1327
+1328
+1329
+1330
+1331
+1332
+1333
+1334
+1335
+1336
+1337
+1338
+1339
+1340
+1341
+1342
+1343
+1344
+1345
+1346
+1347
+1348
+1349
+1350
+1351
+1352
+1353
+1354
+1355
+1356
+1357
+1358
+1359
+1360
+1361
+1362
+1363
+1364
+1365
+1366
+1367
+1368
+1369
+1370
+1371
+1372
+1373
+1374
+1375
+1376
+1377
+1378
+1379
+1380
+1381
+1382
+1383
+1384
+1385
+1386
+1387
+1388
+1389
+1390
+1391
+1392
+1393
+1394
+1395
+1396
+1397
+1398
+1399
+1400
+1401
+1402
+1403
+1404
+1405
+1406
+1407
+1408
+1409
+1410
+1411
+1412
+1413
+1414
+1415
+1416
+1417
+1418
+1419
+1420
+1421
+1422
+1423
+1424
+1425
+1426
+1427
+1428
+1429
+1430
+1431
+1432
+1433
+1434
+1435
+1436
+1437
+1438
+1439
+1440
+1441
+1442
+1443
+1444
+1445
+1446
+1447
+1448
+1449
+1450
+1451
+1452
+1453
+1454
+1455
+1456
+1457
+1458
+1459
+1460
+1461
+1462
+1463
+1464
+1465
+1466
+1467
+1468
+1469
+1470
+1471
+1472
+1473
+1474
+1475
+1476
+1477
+1478
+1479
+1480
+1481
+1482
+1483
+1484
+1485
+1486
+1487
+1488
+1489
+1490
+1491
+1492
+1493
+1494
+1495
+1496
+1497
+1498
+1499
+1500
+1501
+1502
+1503
+1504
+1505
+1506
+1507
+1508
+1509
+1510
+1511
+1512
+1513
+1514
+1515
+1516
+1517
+1518
+1519
+1520
+1521
+1522
+1523
+1524
+1525
+1526
+1527
+1528
+1529
+1530
+1531
+1532
+1533
+1534
+1535
+1536
+1537
+1538
+1539
+1540
+1541
+1542
+1543
+1544
+1545
+1546
+1547
+1548
+1549
+1550
+1551
+1552
+1553
+1554
+1555
+1556
+1557
+1558
+1559
+1560
+1561
+1562
+1563
+1564
+1565
+1566
+1567
+1568
+1569
+1570
+1571
+1572
+1573
+1574
+1575
+1576
+1577
+1578
+1579
+1580
+1581
+1582
+1583
+1584
+1585
+1586
+1587
+1588
+1589
+1590
+1591
+1592
+1593
+1594
+1595
+1596
+1597
+1598
+1599
+1600
+1601
+1602
+1603
+1604
+1605
+1606
+1607
+1608
+1609
+1610
+1611
+1612
+1613
+1614
+1615
+1616
+1617
+1618
+1619
+1620
+1621
+1622
+1623
+1624
+1625
+1626
+1627
+1628
+1629
+1630
+1631
+1632
+1633
+1634
+1635
+1636
+1637
+1638
+1639
+1640
+1641
+1642
+1643
+1644
+1645
+1646
+1647
+1648
+1649
+1650
+1651
+1652
+1653
+1654
+1655
+1656
+1657
+1658
+1659
+1660
+1661
+1662
+1663
+1664
+1665
+1666
+1667
+1668
+1669
+1670
+1671
+1672
+1673
+1674
+1675
+1676
+1677
+1678
+1679
+1680
+1681
+1682
+1683
+1684
+1685
+1686
+1687
+1688
+1689
+1690
+1691
+1692
+1693
+1694
+1695
+1696
+1697
+1698
+1699
+1700
+1701
+1702
+1703
+1704
+1705
+1706
+1707
+1708
+1709
+1710
+1711
+1712
+1713
+1714
+1715
+1716
+1717
+1718
+1719
+1720
+1721
+1722
+1723
+1724
+1725
+1726
+1727
+1728
+1729
+1730
+1731
+1732
+1733
+1734
+1735
+1736
+1737
+1738
+1739
+1740
+1741
+1742
+1743
+1744
+1745
+1746
+1747
+1748
+1749
+1750
+1751
+1752
+1753
+1754
+1755
+1756
+1757
+1758
+1759
+1760
+1761
+1762
+1763
+1764
+1765
+1766
+1767
+1768
+1769
+1770
+1771
+1772
+1773
+1774
+1775
+1776
+1777
+1778
+1779
+1780
+1781
+1782
+1783
+1784
+1785
+1786
+1787
+1788
+1789
+1790
+1791
+1792
+1793
+1794
+1795
+1796
+1797
+1798
+1799
+1800
+1801
+1802
+1803
+1804
+1805
+1806
+1807
+1808
+1809
+1810
+1811
+1812
+1813
+1814
+1815
+1816
+1817
+1818
+1819
+1820
+1821
+1822
+1823
+1824
+1825
+1826
+1827
+1828
+1829
+1830
+1831
+1832
+1833
+1834
+1835
+1836
+1837
+1838
+1839
+1840
+1841
+1842
+1843
+1844
+1845
+1846
+1847
+1848
+1849
+1850
+1851
+1852
+1853
+1854
+1855
+1856
+1857
+1858
+1859
+1860
+1861
+1862
+1863
+1864
+1865
+1866
+1867
+1868
+1869
+1870
+1871
+1872
+1873
+1874
+1875
+1876
+1877
+1878
+1879
class ReduceOptimizer:
+    """
+    A class that optimizes reduce operations in data processing pipelines.
+
+    This optimizer analyzes the input and output of a reduce operation, creates and evaluates
+    multiple reduce plans, and selects the best plan for optimizing the operation's performance.
+
+    Attributes:
+        config (Dict[str, Any]): Configuration dictionary for the optimizer.
+        console (Console): Rich console object for pretty printing.
+        llm_client (LLMClient): Client for interacting with a language model.
+        _run_operation (Callable): Function to run an operation.
+        max_threads (int): Maximum number of threads to use for parallel processing.
+        num_fold_prompts (int): Number of fold prompts to generate.
+        num_samples_in_validation (int): Number of samples to use in validation.
+    """
+
+    def __init__(
+        self,
+        runner,
+        config: Dict[str, Any],
+        console: Console,
+        llm_client: LLMClient,
+        max_threads: int,
+        run_operation: Callable,
+        num_fold_prompts: int = 1,
+        num_samples_in_validation: int = 10,
+        status: Optional[Status] = None,
+    ):
+        """
+        Initialize the ReduceOptimizer.
+
+        Args:
+            config (Dict[str, Any]): Configuration dictionary for the optimizer.
+            console (Console): Rich console object for pretty printing.
+            llm_client (LLMClient): Client for interacting with a language model.
+            max_threads (int): Maximum number of threads to use for parallel processing.
+            run_operation (Callable): Function to run an operation.
+            num_fold_prompts (int, optional): Number of fold prompts to generate. Defaults to 1.
+            num_samples_in_validation (int, optional): Number of samples to use in validation. Defaults to 10.
+        """
+        self.runner = runner
+        self.config = config
+        self.console = console
+        self.llm_client = llm_client
+        self._run_operation = run_operation
+        self.max_threads = max_threads
+        self.num_fold_prompts = num_fold_prompts
+        self.num_samples_in_validation = num_samples_in_validation
+        self.status = status
+
+    def should_optimize_helper(
+        self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]
+    ) -> str:
+        # Check if we're running out of token limits for the reduce prompt
+        model = op_config.get("model", self.config.get("default_model", "gpt-4o-mini"))
+        model_input_context_length = model_cost.get(model, {}).get(
+            "max_input_tokens", 4096
+        )
+
+        # Find the key with the longest value
+        if op_config["reduce_key"] == ["_all"]:
+            sample_key = tuple(["_all"])
+        else:
+            longest_key = max(
+                op_config["reduce_key"], key=lambda k: len(str(input_data[0][k]))
+            )
+            sample_key = tuple(
+                input_data[0][k] if k == longest_key else input_data[0][k]
+                for k in op_config["reduce_key"]
+            )
+
+        # Render the prompt with a sample input
+        prompt_template = Template(op_config["prompt"])
+        sample_prompt = prompt_template.render(
+            reduce_key=dict(zip(op_config["reduce_key"], sample_key)),
+            inputs=[input_data[0]],
+        )
+
+        # Count tokens in the sample prompt
+        prompt_tokens = count_tokens(sample_prompt, model)
+
+        self.console.post_optimizer_status(StageType.SAMPLE_RUN)
+        original_output = self._run_operation(op_config, input_data)
+
+        # Step 1: Synthesize a validator prompt
+        self.console.post_optimizer_status(StageType.SHOULD_OPTIMIZE)
+        validator_prompt = self._generate_validator_prompt(
+            op_config, input_data, original_output
+        )
+
+        # Log the validator prompt
+        self.console.log("[bold]Validator Prompt:[/bold]")
+        self.console.log(validator_prompt)
+        self.console.log("\n")  # Add a newline for better readability
+
+        # Step 2: validate the output
+        validator_inputs = self._create_validation_inputs(
+            input_data, op_config["reduce_key"]
+        )
+        validation_results = self._validate_reduce_output(
+            op_config, validator_inputs, original_output, validator_prompt
+        )
+
+        return validation_results, prompt_tokens, model_input_context_length, model, validator_prompt, original_output
+
+    def should_optimize(self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]) -> Tuple[str, List[Dict[str, Any]], List[Dict[str, Any]]]:
+        validation_results, prompt_tokens, model_input_context_length, model, validator_prompt, original_output = self.should_optimize_helper(op_config, input_data)
+        if prompt_tokens * 1.5 > model_input_context_length:
+            return "The reduce prompt is likely to exceed the token limit for model {model}.", input_data, original_output
+
+        if validation_results.get("needs_improvement", False):
+            return "\n".join(
+                [
+                    f"Issues: {result['issues']} Suggestions: {result['suggestions']}"
+                    for result in validation_results["validation_results"]
+                ]
+            ), input_data, original_output
+        else:
+            return "", input_data, original_output
+
+    def optimize(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        level: int = 1,
+    ) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]:
+        """
+        Optimize the reduce operation based on the given configuration and input data.
+
+        This method performs the following steps:
+        1. Run the original operation
+        2. Generate a validator prompt
+        3. Validate the output
+        4. If improvement is needed:
+           a. Evaluate if decomposition is beneficial
+           b. If decomposition is beneficial, recursively optimize each sub-operation
+           c. If not, proceed with single operation optimization
+        5. Run the optimized operation(s)
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+
+        Returns:
+            Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]: A tuple containing the list of optimized configurations
+            and the list of outputs from the optimized operation(s), and the cost of the operation due to synthesizing any resolve operations.
+        """
+        validation_results, prompt_tokens, model_input_context_length, model, validator_prompt, original_output = self.should_optimize_helper(op_config, input_data)
+
+        add_map_op = False
+        if prompt_tokens * 2 > model_input_context_length:
+            add_map_op = True
+            self.console.log(
+                f"[yellow]Warning: The reduce prompt exceeds the token limit for model {model}. "
+                f"Token count: {prompt_tokens}, Limit: {model_input_context_length}. "
+                f"Add a map operation to the pipeline.[/yellow]"
+            )
+
+        # # Also query an agent to look at a sample of the inputs and see if they think a map operation would be helpful
+        # preprocessing_steps = ""
+        # should_use_map, preprocessing_steps = self._should_use_map(
+        #     op_config, input_data
+        # )
+        # if should_use_map or add_map_op:
+        #     # Synthesize a map operation
+        #     map_prompt, map_output_schema = self._synthesize_map_operation(
+        #         op_config, preprocessing_steps, input_data
+        #     )
+        #     # Change the reduce operation prompt to use the map schema
+        #     new_reduce_prompt = self._change_reduce_prompt_to_use_map_schema(
+        #         op_config["prompt"], map_output_schema
+        #     )
+        #     op_config["prompt"] = new_reduce_prompt
+
+        #     # Return unoptimized map and reduce operations
+        #     return [map_prompt, op_config], input_data, 0.0
+
+
+        # Print the validation results
+        self.console.log("[bold]Validation Results on Initial Sample:[/bold]")
+        if validation_results["needs_improvement"] or self.config.get("optimizer_config", {}).get("force_decompose", False):
+            self.console.post_optimizer_rationale(
+                should_optimize=True,
+                rationale= "\n".join(
+                    [
+                        f"Issues: {result['issues']} Suggestions: {result['suggestions']}"
+                        for result in validation_results["validation_results"]
+                    ]
+                ),
+                validator_prompt=validator_prompt,
+            )
+            self.console.log(
+                "\n".join(
+                    [
+                        f"Issues: {result['issues']} Suggestions: {result['suggestions']}"
+                        for result in validation_results["validation_results"]
+                    ]
+                )
+            )
+
+            # Step 3: Evaluate if decomposition is beneficial
+            decomposition_result = self._evaluate_decomposition(
+                op_config, input_data, level
+            )
+
+            if decomposition_result["should_decompose"]:
+                return self._optimize_decomposed_reduce(
+                    decomposition_result, op_config, input_data, level
+                )
+
+            return self._optimize_single_reduce(op_config, input_data, validator_prompt)
+        else:
+            self.console.log(f"No improvements identified; {validation_results}.")
+            self.console.post_optimizer_rationale(
+                should_optimize=False,
+                rationale="No improvements identified; no optimization recommended.",
+                validator_prompt=validator_prompt,
+            )
+            return [op_config], original_output, 0.0
+
+    def _should_use_map(
+        self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]
+    ) -> Tuple[bool, str]:
+        """
+        Determine if a map operation should be used based on the input data.
+        """
+        # Sample a random input item
+        sample_input = random.choice(input_data)
+
+        # Format the prompt with the sample input
+        prompt_template = Template(op_config["prompt"])
+        formatted_prompt = prompt_template.render(
+            reduce_key=dict(
+                zip(op_config["reduce_key"], sample_input[op_config["reduce_key"]])
+            ),
+            inputs=[sample_input],
+        )
+
+        # Prepare the message for the LLM
+        messages = [{"role": "user", "content": formatted_prompt}]
+
+        # Truncate the messages to fit the model's context window
+        truncated_messages = truncate_messages(
+            messages, self.config.get("model", self.default_model)
+        )
+
+        # Query the LLM for preprocessing suggestions
+        preprocessing_prompt = (
+            "Based on the following reduce operation prompt, should we do any preprocessing on the input data? "
+            "Consider if we need to remove unnecessary context, or logically construct an output that will help in the task. "
+            "If preprocessing would be beneficial, explain why and suggest specific steps. If not, explain why preprocessing isn't necessary.\n\n"
+            f"Reduce operation prompt:\n{truncated_messages[0]['content']}"
+        )
+
+        preprocessing_response = self.llm_client.generate(
+            model=self.config.get("model", self.default_model),
+            messages=[{"role": "user", "content": preprocessing_prompt}],
+            response_format={
+                "type": "json_object",
+                "schema": {
+                    "type": "object",
+                    "properties": {
+                        "preprocessing_needed": {"type": "boolean"},
+                        "rationale": {"type": "string"},
+                        "suggested_steps": {"type": "string"},
+                    },
+                    "required": [
+                        "preprocessing_needed",
+                        "rationale",
+                        "suggested_steps",
+                    ],
+                },
+            },
+        )
+
+        preprocessing_result = preprocessing_response.choices[0].message.content
+
+        should_preprocess = preprocessing_result["preprocessing_needed"]
+        preprocessing_rationale = preprocessing_result["rationale"]
+
+        self.console.log(f"[bold]Map-Reduce Decomposition Analysis:[/bold]")
+        self.console.log(f"Should write a map operation: {should_preprocess}")
+        self.console.log(f"Rationale: {preprocessing_rationale}")
+
+        if should_preprocess:
+            self.console.log(
+                f"Suggested steps: {preprocessing_result['suggested_steps']}"
+            )
+
+        return should_preprocess, preprocessing_result["suggested_steps"]
+
+    def _optimize_single_reduce(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        validator_prompt: str,
+    ) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]:
+        """
+        Optimize a single reduce operation.
+
+        This method performs the following steps:
+        1. Determine and configure value sampling
+        2. Determine if the reduce operation is associative
+        3. Create and evaluate multiple reduce plans
+        4. Run the best reduce plan
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+            validator_prompt (str): The validator prompt for evaluating reduce plans.
+
+        Returns:
+            Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]: A tuple containing a single-item list with the optimized configuration
+            and a single-item list with the output from the optimized operation, and the cost of the operation due to synthesizing any resolve operations.
+        """
+        # Step 1: Determine and configure value sampling (TODO: re-enable this when the agent is more reliable)
+        # value_sampling_config = self._determine_value_sampling(op_config, input_data)
+        # if value_sampling_config["enabled"]:
+        #     op_config["value_sampling"] = value_sampling_config
+        #     self.console.log("[bold]Value Sampling Configuration:[/bold]")
+        #     self.console.log(json.dumps(value_sampling_config, indent=2))
+
+        # Step 2: Determine if the reduce operation is associative
+        is_associative = self._is_associative(op_config, input_data)
+
+        # Step 3: Create and evaluate multiple reduce plans
+        self.console.post_optimizer_status(StageType.CANDIDATE_PLANS)
+        self.console.log("[bold magenta]Generating batched plans...[/bold magenta]")
+        reduce_plans = self._create_reduce_plans(op_config, input_data, is_associative)
+
+        # Create gleaning plans
+        self.console.log("[bold magenta]Generating gleaning plans...[/bold magenta]")
+        gleaning_plans = self._generate_gleaning_plans(reduce_plans, validator_prompt)
+
+        self.console.log("[bold magenta]Evaluating plans...[/bold magenta]")
+        self.console.post_optimizer_status(StageType.EVALUATION_RESULTS)
+        best_plan = self._evaluate_reduce_plans(
+            op_config, reduce_plans + gleaning_plans, input_data, validator_prompt
+        )
+
+        # Step 4: Run the best reduce plan
+        optimized_output = self._run_operation(best_plan, input_data)
+        self.console.post_optimizer_status(StageType.END)
+
+        return [best_plan], optimized_output, 0.0
+
+    def _generate_gleaning_plans(
+        self,
+        plans: List[Dict[str, Any]],
+        validation_prompt: str,
+    ) -> List[Dict[str, Any]]:
+        """
+        Generate plans that use gleaning for the given operation.
+
+        Gleaning involves iteratively refining the output of an operation
+        based on validation feedback. This method creates plans with different
+        numbers of gleaning rounds.
+
+        Args:
+            plans (List[Dict[str, Any]]): The list of plans to use for gleaning.
+            validation_prompt (str): The prompt used for validating the operation's output.
+
+        Returns:
+            Dict[str, List[Dict[str, Any]]]: A dictionary of gleaning plans, where each key
+            is a plan name and each value is a list containing a single operation configuration
+            with gleaning parameters.
+
+        """
+        # Generate an op with gleaning num_rounds and validation_prompt
+        gleaning_plans = []
+        gleaning_rounds = [1]
+        biggest_batch_size = max([plan["fold_batch_size"] for plan in plans])
+        for plan in plans:
+            if plan["fold_batch_size"] != biggest_batch_size:
+                continue
+            for gleaning_round in gleaning_rounds:
+                plan_copy = copy.deepcopy(plan)
+                plan_copy["gleaning"] = {
+                    "num_rounds": gleaning_round,
+                    "validation_prompt": validation_prompt,
+                }
+                plan_name = f"gleaning_{gleaning_round}_rounds_{plan['name']}"
+                plan_copy["name"] = plan_name
+                gleaning_plans.append(plan_copy)
+        return gleaning_plans
+
+    def _optimize_decomposed_reduce(
+        self,
+        decomposition_result: Dict[str, Any],
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        level: int,
+    ) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]:
+        """
+        Optimize a decomposed reduce operation.
+
+        This method performs the following steps:
+        1. Group the input data by the sub-group key.
+        2. Optimize the first reduce operation.
+        3. Run the optimized first reduce operation on all groups.
+        4. Optimize the second reduce operation using the results of the first.
+        5. Run the optimized second reduce operation.
+
+        Args:
+            decomposition_result (Dict[str, Any]): The result of the decomposition evaluation.
+            op_config (Dict[str, Any]): The original reduce operation configuration.
+            input_data (List[Dict[str, Any]]): The input data for the reduce operation.
+            level (int): The current level of decomposition.
+        Returns:
+            Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]: A tuple containing the list of optimized configurations
+            for both reduce operations and the final output of the second reduce operation, and the cost of the operation due to synthesizing any resolve operations.
+        """
+        sub_group_key = decomposition_result["sub_group_key"]
+        first_reduce_prompt = decomposition_result["first_reduce_prompt"]
+        second_reduce_prompt = decomposition_result["second_reduce_prompt"]
+        pipeline = []
+        all_cost = 0.0
+
+        first_reduce_config = op_config.copy()
+        first_reduce_config["prompt"] = first_reduce_prompt
+        if isinstance(op_config["reduce_key"], list):
+            first_reduce_config["reduce_key"] = [sub_group_key] + op_config[
+                "reduce_key"
+            ]
+        else:
+            first_reduce_config["reduce_key"] = [sub_group_key, op_config["reduce_key"]]
+        first_reduce_config["pass_through"] = True
+
+        if first_reduce_config.get("synthesize_resolve", True):
+            resolve_config = {
+                "type": "resolve",
+                "empty": True,
+                "embedding_model": "text-embedding-3-small",
+                "resolution_model": self.config.get("default_model", "gpt-4o-mini"),
+                "comparison_model": self.config.get("default_model", "gpt-4o-mini"),
+                "_intermediates": {
+                    "map_prompt": op_config.get("_intermediates", {}).get(
+                        "last_map_prompt"
+                    ),
+                    "reduce_key": first_reduce_config["reduce_key"],
+                },
+            }
+            optimized_resolve_config, resolve_cost = JoinOptimizer(
+                self.config,
+                resolve_config,
+                self.console,
+                self.llm_client,
+                self.max_threads,
+            ).optimize_resolve(input_data)
+            all_cost += resolve_cost
+
+            if not optimized_resolve_config.get("empty", False):
+                # Add this to the pipeline
+                pipeline += [optimized_resolve_config]
+
+                # Run the resolver
+                optimized_output = self._run_operation(
+                    optimized_resolve_config, input_data
+                )
+                input_data = optimized_output
+
+        first_optimized_configs, first_outputs, first_cost = self.optimize(
+            first_reduce_config, input_data, level + 1
+        )
+        pipeline += first_optimized_configs
+        all_cost += first_cost
+
+        # Optimize second reduce operation
+        second_reduce_config = op_config.copy()
+        second_reduce_config["prompt"] = second_reduce_prompt
+        second_reduce_config["pass_through"] = True
+
+        second_optimized_configs, second_outputs, second_cost = self.optimize(
+            second_reduce_config, first_outputs, level + 1
+        )
+
+        # Combine optimized configs and return with final output
+        pipeline += second_optimized_configs
+        all_cost += second_cost
+
+        return pipeline, second_outputs, all_cost
+
+    def _evaluate_decomposition(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        level: int = 1,
+    ) -> Dict[str, Any]:
+        """
+        Evaluate whether decomposing the reduce operation would be beneficial.
+
+        This method first determines if decomposition would be helpful, and if so,
+        it then determines the sub-group key and prompts for the decomposed operations.
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+            level (int): The current level of decomposition.
+
+        Returns:
+            Dict[str, Any]: A dictionary containing the decomposition decision and details.
+        """
+        should_decompose = self._should_decompose(op_config, input_data, level)
+
+        # Log the decomposition decision
+        if should_decompose["should_decompose"]:
+            self.console.log(
+                f"[bold green]Decomposition recommended:[/bold green] {should_decompose['explanation']}"
+            )
+        else:
+            self.console.log(
+                f"[bold yellow]Decomposition not recommended:[/bold yellow] {should_decompose['explanation']}"
+            )
+
+        # Return early if decomposition is not recommended
+        if not should_decompose["should_decompose"]:
+            return should_decompose
+
+        # Temporarily stop the status
+        if self.status:
+            self.status.stop()
+
+        # Ask user if they agree with the decomposition assessment
+        user_agrees = Confirm.ask(
+            f"Do you agree with the decomposition assessment? "
+            f"[bold]{'Recommended' if should_decompose['should_decompose'] else 'Not recommended'}[/bold]",
+            console=self.console,
+        )
+
+        # If user disagrees, invert the decomposition decision
+        if not user_agrees:
+            should_decompose["should_decompose"] = not should_decompose[
+                "should_decompose"
+            ]
+            should_decompose["explanation"] = (
+                "User disagreed with the initial assessment."
+            )
+
+        # Restart the status
+        if self.status:
+            self.status.start()
+
+        # Return if decomposition is not recommended
+        if not should_decompose["should_decompose"]:
+            return should_decompose
+
+        decomposition_details = self._get_decomposition_details(op_config, input_data)
+        result = {**should_decompose, **decomposition_details}
+        if decomposition_details["sub_group_key"] in op_config["reduce_key"]:
+            result["should_decompose"] = False
+            result[
+                "explanation"
+            ] += " However, the suggested sub-group key is already part of the current reduce key(s), so decomposition is not recommended."
+            result["sub_group_key"] = ""
+
+        return result
+
+    def _should_decompose(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        level: int = 1,
+    ) -> Dict[str, Any]:
+        """
+        Determine if decomposing the reduce operation would be beneficial.
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+            level (int): The current level of decomposition.
+
+        Returns:
+            Dict[str, Any]: A dictionary containing the decomposition decision and explanation.
+        """
+        # TODO: we have not enabled recursive decomposition yet
+        if level > 1 and not op_config.get("recursively_optimize", False):
+            return {
+                "should_decompose": False,
+                "explanation": "Recursive decomposition is not enabled.",
+            }
+
+        system_prompt = (
+            "You are an AI assistant tasked with optimizing data processing pipelines."
+        )
+
+        # Sample a subset of input data for analysis
+        sample_size = min(10, len(input_data))
+        sample_input = random.sample(input_data, sample_size)
+
+        # Get all keys from the input data
+        all_keys = set().union(*(item.keys() for item in sample_input))
+        reduce_key = op_config["reduce_key"]
+        reduce_keys = [reduce_key] if isinstance(reduce_key, str) else reduce_key
+        other_keys = [key for key in all_keys if key not in reduce_keys]
+
+        # See if there's an input schema and constrain the sample_input to that schema
+        input_schema = op_config.get("input", {}).get("schema", {})
+        if input_schema:
+            sample_input = [
+                {key: item[key] for key in input_schema} for item in sample_input
+            ]
+
+        # Create a sample of values for other keys
+        sample_values = {
+            key: list(set(str(item.get(key))[:50] for item in sample_input))[:5]
+            for key in other_keys
+        }
+
+        prompt = f"""Analyze the following reduce operation and determine if it should be decomposed into two reduce operations chained together:
+
+        Reduce Operation Prompt:
+        ```
+        {op_config['prompt']}
+        ```
+
+        Current Reduce Key(s): {reduce_keys}
+        Other Available Keys: {', '.join(other_keys)}
+
+        Sample values for other keys:
+        {json.dumps(sample_values, indent=2)}
+
+        Based on this information, determine if it would be beneficial to decompose this reduce operation into a sub-reduce operation followed by a final reduce operation. Consider ALL of the following:
+
+        1. Is there a natural hierarchy in the data (e.g., country -> state -> city) among the other available keys, with a key at a finer level of granularity than the current reduce key(s)?
+        2. Are the current reduce key(s) some form of ID, and are there many different types of inputs for that ID among the other available keys?
+        3. Does the prompt implicitly ask for sub-grouping based on the other available keys (e.g., "summarize policies by state, then by country")?
+        4. Would splitting the operation improve accuracy (i.e., make sure information isn't lost when reducing)?
+        5. Are all the keys of the potential hierarchy provided in the other available keys? If not, we should not decompose.
+        6. Importantly, do not suggest decomposition using any key that is already part of the current reduce key(s). We are looking for a new key from the other available keys to use for sub-grouping.
+        7. Do not suggest keys that don't contain meaningful information (e.g., id-related keys).
+
+        Provide your analysis in the following format:
+        """
+
+        parameters = {
+            "type": "object",
+            "properties": {
+                "should_decompose": {"type": "boolean"},
+                "explanation": {"type": "string"},
+            },
+            "required": ["should_decompose", "explanation"],
+        }
+
+        response = self.llm_client.generate(
+            [{"role": "user", "content": prompt}],
+            system_prompt,
+            parameters,
+        )
+        return json.loads(response.choices[0].message.content)
+
+    def _get_decomposition_details(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+    ) -> Dict[str, Any]:
+        """
+        Determine the sub-group key and prompts for decomposed reduce operations.
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+
+        Returns:
+            Dict[str, Any]: A dictionary containing the sub-group key and prompts for decomposed operations.
+        """
+        system_prompt = (
+            "You are an AI assistant tasked with optimizing data processing pipelines."
+        )
+
+        # Sample a subset of input data for analysis
+        sample_size = min(10, len(input_data))
+        sample_input = random.sample(input_data, sample_size)
+
+        # Get all keys from the input data
+        all_keys = set().union(*(item.keys() for item in sample_input))
+        reduce_key = op_config["reduce_key"]
+        reduce_keys = [reduce_key] if isinstance(reduce_key, str) else reduce_key
+        other_keys = [key for key in all_keys if key not in reduce_keys]
+
+        prompt = f"""Given that we've decided to decompose the following reduce operation, suggest a two-step reduce process:
+
+        Reduce Operation Prompt:
+        ```
+        {op_config['prompt']}
+        ```
+
+        Reduce Key(s): {reduce_key}
+        Other Keys: {', '.join(other_keys)}
+
+        Provide the following:
+        1. A sub-group key to use for the first reduce operation
+        2. A prompt for the first reduce operation
+        3. A prompt for the second (final) reduce operation
+
+        For the reduce operation prompts, you should only minimally modify the original prompt. The prompts should be Jinja templates, and the only variables they can access are the `reduce_key` and `inputs` variables.
+
+        Provide your suggestions in the following format:
+        """
+
+        parameters = {
+            "type": "object",
+            "properties": {
+                "sub_group_key": {"type": "string"},
+                "first_reduce_prompt": {"type": "string"},
+                "second_reduce_prompt": {"type": "string"},
+            },
+            "required": [
+                "sub_group_key",
+                "first_reduce_prompt",
+                "second_reduce_prompt",
+            ],
+        }
+
+        response = self.llm_client.generate(
+            [{"role": "user", "content": prompt}],
+            system_prompt,
+            parameters,
+        )
+        return json.loads(response.choices[0].message.content)
+
+    def _determine_value_sampling(
+        self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]
+    ) -> Dict[str, Any]:
+        """
+        Determine whether value sampling should be enabled and configure its parameters.
+        """
+        system_prompt = (
+            "You are an AI assistant helping to optimize data processing pipelines."
+        )
+
+        # Sample a subset of input data for analysis
+        sample_size = min(100, len(input_data))
+        sample_input = random.sample(input_data, sample_size)
+
+        prompt = f"""
+        Analyze the following reduce operation and determine if value sampling should be enabled:
+
+        Reduce Operation Prompt:
+        {op_config['prompt']}
+
+        Sample Input Data (first 2 items):
+        {json.dumps(sample_input[:2], indent=2)}
+
+        Value sampling is appropriate for reduce operations that don't need to look at all the values for each key to produce a good result, such as generic summarization tasks.
+
+        Based on the reduce operation prompt and the sample input data, determine if value sampling should be enabled.
+        Answer with 'yes' if value sampling should be enabled or 'no' if it should not be enabled. Explain your reasoning briefly.
+        """
+
+        parameters = {
+            "type": "object",
+            "properties": {
+                "enable_sampling": {"type": "boolean"},
+                "explanation": {"type": "string"},
+            },
+            "required": ["enable_sampling", "explanation"],
+        }
+
+        response = self.llm_client.generate(
+            [{"role": "user", "content": prompt}],
+            system_prompt,
+            parameters,
+        )
+        result = json.loads(response.choices[0].message.content)
+
+        if not result["enable_sampling"]:
+            return {"enabled": False}
+
+        # Print the explanation for enabling value sampling
+        self.console.log(f"Value sampling enabled: {result['explanation']}")
+
+        # Determine sampling method
+        prompt = f"""
+        We are optimizing a reduce operation in a data processing pipeline. The reduce operation is defined by the following prompt:
+
+        Reduce Operation Prompt:
+        {op_config['prompt']}
+
+        Sample Input Data (first 2 items):
+        {json.dumps(sample_input[:2], indent=2)}
+
+        We have determined that value sampling should be enabled for this reduce operation. Value sampling is a technique used to process only a subset of the input data for each reduce key, rather than processing all items. This can significantly reduce processing time and costs for very large datasets, especially when the reduce operation doesn't require looking at every single item to produce a good result (e.g., summarization tasks).
+
+        Now we need to choose the most appropriate sampling method. The available methods are:
+
+        1. "random": Randomly select a subset of values.
+        Example: In a customer review analysis task, randomly selecting a subset of reviews to summarize the overall sentiment.
+
+        2. "cluster": Use K-means clustering to select representative samples.
+        Example: In a document categorization task, clustering documents based on their content and selecting representative documents from each cluster to determine the overall categories.
+
+        3. "sem_sim": Use semantic similarity to select the most relevant samples to a query text.
+        Example: In a news article summarization task, selecting articles that are semantically similar to a query like "Major economic events of {{reduce_key}}" to produce a focused summary.
+
+        Based on the reduce operation prompt, the nature of the task, and the sample input data, which sampling method would be most appropriate?
+
+        Provide your answer as either "random", "cluster", or "sem_sim", and explain your reasoning in detail. Consider the following in your explanation:
+        - The nature of the reduce task (e.g., summarization, aggregation, analysis)
+        - The structure and content of the input data
+        - The potential benefits and drawbacks of each sampling method for this specific task
+        """
+
+        parameters = {
+            "type": "object",
+            "properties": {
+                "method": {"type": "string", "enum": ["random", "cluster", "sem_sim"]},
+                "explanation": {"type": "string"},
+            },
+            "required": ["method", "explanation"],
+        }
+
+        response = self.llm_client.generate(
+            [{"role": "user", "content": prompt}],
+            system_prompt,
+            parameters,
+        )
+        result = json.loads(response.choices[0].message.content)
+        method = result["method"]
+
+        value_sampling_config = {
+            "enabled": True,
+            "method": method,
+            "sample_size": 100,  # Default sample size
+            "embedding_model": "text-embedding-3-small",
+        }
+
+        if method in ["cluster", "sem_sim"]:
+            # Determine embedding keys
+            prompt = f"""
+            For the {method} sampling method, we need to determine which keys from the input data should be used for generating embeddings.
+
+            Input data keys:
+            {', '.join(sample_input[0].keys())}
+
+            Sample Input Data:
+            {json.dumps(sample_input[0], indent=2)[:1000]}...
+
+            Based on the reduce operation prompt and the sample input data, which keys should be used for generating embeddings? Use keys that will create meaningful embeddings (i.e., not id-related keys).
+            Provide your answer as a list of key names that is a subset of the input data keys. You should pick only the 1-3 keys that are necessary for generating meaningful embeddings, that have relatively short values.
+            """
+
+            parameters = {
+                "type": "object",
+                "properties": {
+                    "embedding_keys": {"type": "array", "items": {"type": "string"}},
+                    "explanation": {"type": "string"},
+                },
+                "required": ["embedding_keys", "explanation"],
+            }
+
+            response = self.llm_client.generate(
+                [{"role": "user", "content": prompt}],
+                system_prompt,
+                parameters,
+            )
+            result = json.loads(response.choices[0].message.content)
+            # TODO: validate that these exist
+            embedding_keys = result["embedding_keys"]
+            for key in result["embedding_keys"]:
+                if key not in sample_input[0]:
+                    embedding_keys.remove(key)
+
+            if not embedding_keys:
+                # Select the reduce key
+                self.console.log(
+                    "No embedding keys found, selecting reduce key for embedding key"
+                )
+                embedding_keys = (
+                    op_config["reduce_key"]
+                    if isinstance(op_config["reduce_key"], list)
+                    else [op_config["reduce_key"]]
+                )
+
+            value_sampling_config["embedding_keys"] = embedding_keys
+
+        if method == "sem_sim":
+            # Determine query text
+            prompt = f"""
+            For the semantic similarity (sem_sim) sampling method, we need to determine the query text to compare against when selecting samples.
+
+            Reduce Operation Prompt:
+            {op_config['prompt']}
+
+            The query text should be a Jinja template with access to the `reduce_key` variable.
+            Based on the reduce operation prompt, what would be an appropriate query text for selecting relevant samples?
+            """
+
+            parameters = {
+                "type": "object",
+                "properties": {
+                    "query_text": {"type": "string"},
+                    "explanation": {"type": "string"},
+                },
+                "required": ["query_text", "explanation"],
+            }
+
+            response = self.llm_client.generate(
+                [{"role": "user", "content": prompt}],
+                system_prompt,
+                parameters,
+            )
+            result = json.loads(response.choices[0].message.content)
+            value_sampling_config["query_text"] = result["query_text"]
+
+        return value_sampling_config
+
+    def _is_associative(
+        self, op_config: Dict[str, Any], input_data: List[Dict[str, Any]]
+    ) -> bool:
+        """
+        Determine if the reduce operation is associative.
+
+        This method analyzes the reduce operation configuration and a sample of the input data
+        to determine if the operation is associative (i.e., the order of combining elements
+        doesn't affect the final result).
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+
+        Returns:
+            bool: True if the operation is determined to be associative, False otherwise.
+        """
+        system_prompt = (
+            "You are an AI assistant helping to optimize data processing pipelines."
+        )
+
+        # Sample a subset of input data for analysis
+        sample_size = min(5, len(input_data))
+        sample_input = random.sample(input_data, sample_size)
+
+        prompt = f"""
+        Analyze the following reduce operation and determine if it is associative:
+
+        Reduce Operation Prompt:
+        {op_config['prompt']}
+
+        Sample Input Data:
+        {json.dumps(sample_input, indent=2)[:1000]}...
+
+        Based on the reduce operation prompt, determine whether the order in which we process data matters.
+        Answer with 'yes' if order matters or 'no' if order doesn't matter.
+        Explain your reasoning briefly.
+
+        For example:
+        - Merging extracted key-value pairs from documents does not require order: combining {{"name": "John", "age": 30}} with {{"city": "New York", "job": "Engineer"}} yields the same result regardless of order
+        - Generating a timeline of events requires order: the order of events matters for maintaining chronological accuracy.
+
+        Consider these examples when determining whether the order in which we process data matters. You might also have to consider the specific data.
+        """
+
+        parameters = {
+            "type": "object",
+            "properties": {
+                "order_matters": {"type": "boolean"},
+                "explanation": {"type": "string"},
+            },
+            "required": ["order_matters", "explanation"],
+        }
+
+        response = self.llm_client.generate(
+            [{"role": "user", "content": prompt}],
+            system_prompt,
+            parameters,
+        )
+        result = json.loads(response.choices[0].message.content)
+        result["is_associative"] = not result["order_matters"]
+
+        self.console.log(
+            f"[yellow]Reduce operation {'is associative' if result['is_associative'] else 'is not associative'}.[/yellow] Analysis: {result['explanation']}"
+        )
+        return result["is_associative"]
+
+    def _generate_validator_prompt(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        original_output: List[Dict[str, Any]],
+    ) -> str:
+        """
+        Generate a custom validator prompt for assessing the quality of the reduce operation output.
+
+        This method creates a prompt that will be used to validate the output of the reduce operation.
+        It includes specific questions about the quality and completeness of the output.
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+            original_output (List[Dict[str, Any]]): Original output of the reduce operation.
+
+        Returns:
+            str: A custom validator prompt as a string.
+        """
+        system_prompt = "You are an AI assistant tasked with creating custom validation prompts for reduce operations in data processing pipelines."
+
+        sample_input = random.choice(input_data)
+        input_keys = op_config.get("input", {}).get("schema", {})
+        if input_keys:
+            sample_input = {k: sample_input[k] for k in input_keys}
+
+        reduce_key = op_config.get("reduce_key")
+        if reduce_key and original_output:
+            if isinstance(reduce_key, list):
+                key = next(
+                    (
+                        tuple(item[k] for k in reduce_key)
+                        for item in original_output
+                        if all(k in item for k in reduce_key)
+                    ),
+                    tuple(None for _ in reduce_key),
+                )
+                sample_output = next(
+                    (
+                        item
+                        for item in original_output
+                        if all(item.get(k) == v for k, v in zip(reduce_key, key))
+                    ),
+                    {},
+                )
+            else:
+                key = next(
+                    (
+                        item[reduce_key]
+                        for item in original_output
+                        if reduce_key in item
+                    ),
+                    None,
+                )
+                sample_output = next(
+                    (item for item in original_output if item.get(reduce_key) == key),
+                    {},
+                )
+        else:
+            sample_output = original_output[0] if original_output else {}
+
+        output_keys = op_config.get("output", {}).get("schema", {})
+        sample_output = {k: sample_output[k] for k in output_keys}
+
+        prompt = f"""
+        Analyze the following reduce operation and its input/output:
+
+        Reduce Operation Prompt:
+        {op_config["prompt"]}
+
+        Sample Input (just one item):
+        {json.dumps(sample_input, indent=2)}
+
+        Sample Output:
+        {json.dumps(sample_output, indent=2)}
+
+        Create a custom validator prompt that will assess how well the reduce operation performed its intended task. The prompt should ask specific 2-3 questions about the quality of the output, such as:
+        1. Does the output accurately reflect the aggregation method specified in the task? For example, if finding anomalies, are the identified anomalies actually anomalies?
+        2. Are there any missing fields, unexpected null values, or data type mismatches in the output compared to the expected schema?
+        3. Does the output maintain the key information from the input while appropriately condensing or summarizing it? For instance, in a text summarization task, are the main points preserved?
+        4. How well does the output adhere to any specific formatting requirements mentioned in the original prompt, such as character limits for summaries or specific data types for aggregated values?
+
+        Note that the output may reflect more than just the input provided, since we only provide a one-item sample input. Provide your response as a single string containing the custom validator prompt. The prompt should be tailored to the task and avoid generic criteria. The prompt should not reference a specific value in the sample input, but rather a general property.
+        """
+
+        parameters = {
+            "type": "object",
+            "properties": {"validator_prompt": {"type": "string"}},
+            "required": ["validator_prompt"],
+        }
+
+        response = self.llm_client.generate(
+            [{"role": "user", "content": prompt}],
+            system_prompt,
+            parameters,
+        )
+        return json.loads(response.choices[0].message.content)["validator_prompt"]
+
+    def _validate_reduce_output(
+        self,
+        op_config: Dict[str, Any],
+        validation_inputs: Dict[Any, List[Dict[str, Any]]],
+        output_data: List[Dict[str, Any]],
+        validator_prompt: str,
+    ) -> Dict[str, Any]:
+        """
+        Validate the output of the reduce operation using the generated validator prompt.
+
+        This method assesses the quality of the reduce operation output by applying the validator prompt
+        to multiple samples of the input and output data.
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            validation_inputs (Dict[Any, List[Dict[str, Any]]]): Validation inputs for the reduce operation.
+            output_data (List[Dict[str, Any]]): Output data from the reduce operation.
+            validator_prompt (str): The validator prompt generated earlier.
+
+        Returns:
+            Dict[str, Any]: A dictionary containing validation results and a flag indicating if improvement is needed.
+        """
+        system_prompt = "You are an AI assistant tasked with validating the output of reduce operations in data processing pipelines."
+
+        validation_results = []
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            futures = []
+            for reduce_key, inputs in validation_inputs.items():
+                if op_config["reduce_key"] == ["_all"] or op_config["reduce_key"] == "_all":
+                    sample_output = output_data[0]
+                elif isinstance(op_config["reduce_key"], list):
+                    sample_output = next(
+                        (
+                            item
+                            for item in output_data
+                            if all(
+                                item[key] == reduce_key[i]
+                                for i, key in enumerate(op_config["reduce_key"])
+                            )
+                        ),
+                        None,
+                    )
+                else:
+                    sample_output = next(
+                        (
+                            item
+                            for item in output_data
+                            if item[op_config["reduce_key"]] == reduce_key
+                        ),
+                        None,
+                    )
+
+                if sample_output is None:
+                    self.console.log(
+                        f"Warning: No output found for reduce key {reduce_key}"
+                    )
+                    continue
+
+                input_str = json.dumps(inputs, indent=2)
+                # truncate input_str to 40,000 words
+                input_str = input_str.split()[:40000]
+                input_str = " ".join(input_str) + "..."
+
+                prompt = f"""{validator_prompt}
+
+                Reduce Operation Task:
+                {op_config["prompt"]}
+
+                Input Data Samples:
+                {input_str}
+
+                Output Data Sample:
+                {json.dumps(sample_output, indent=2)}
+
+                Based on the validator prompt and the input/output samples, assess the quality (e.g., correctness, completeness) of the reduce operation output.
+                Provide your assessment in the following format:
+                """
+
+                parameters = {
+                    "type": "object",
+                    "properties": {
+                        "is_correct": {"type": "boolean"},
+                        "issues": {"type": "array", "items": {"type": "string"}},
+                        "suggestions": {"type": "array", "items": {"type": "string"}},
+                    },
+                    "required": ["is_correct", "issues", "suggestions"],
+                }
+
+                futures.append(
+                    executor.submit(
+                        self.llm_client.generate,
+                        [{"role": "user", "content": prompt}],
+                        system_prompt,
+                        parameters,
+                    )
+                )
+
+            for future, (reduce_key, inputs) in zip(futures, validation_inputs.items()):
+                response = future.result()
+                result = json.loads(response.choices[0].message.content)
+                validation_results.append(result)
+
+        # Determine if optimization is needed based on validation results
+        invalid_count = sum(
+            1 for result in validation_results if not result["is_correct"]
+        )
+        needs_improvement = invalid_count > 1 or (
+            invalid_count == 1 and len(validation_results) == 1
+        )
+
+        return {
+            "needs_improvement": needs_improvement,
+            "validation_results": validation_results,
+        }
+
+    def _create_validation_inputs(
+        self, input_data: List[Dict[str, Any]], reduce_key: Union[str, List[str]]
+    ) -> Dict[Any, List[Dict[str, Any]]]:
+        # Group input data by reduce_key
+        grouped_data = {}
+        if reduce_key == ["_all"]:
+            # Put all data in one group under a single key
+            grouped_data[("_all",)] = input_data
+        else:
+            # Group by reduce key(s) as before
+            for item in input_data:
+                if isinstance(reduce_key, list):
+                    key = tuple(item[k] for k in reduce_key)
+                else:
+                    key = item[reduce_key]
+                if key not in grouped_data:
+                    grouped_data[key] = []
+                grouped_data[key].append(item)
+
+        # Select a fixed number of reduce keys
+        selected_keys = random.sample(
+            list(grouped_data.keys()),
+            min(self.num_samples_in_validation, len(grouped_data)),
+        )
+
+        # Create a new dict with only the selected keys
+        validation_inputs = {key: grouped_data[key] for key in selected_keys}
+
+        return validation_inputs
+
+    def _create_reduce_plans(
+        self,
+        op_config: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        is_associative: bool,
+    ) -> List[Dict[str, Any]]:
+        """
+        Create multiple reduce plans based on the input data and operation configuration.
+
+        This method generates various reduce plans by varying batch sizes and fold prompts.
+        It takes into account the LLM's context window size to determine appropriate batch sizes.
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+            is_associative (bool): Flag indicating whether the reduce operation is associative.
+
+        Returns:
+            List[Dict[str, Any]]: A list of reduce plans, each with different batch sizes and fold prompts.
+        """
+        model = op_config.get("model", "gpt-4o-mini")
+        model_input_context_length = model_cost.get(model, {}).get(
+            "max_input_tokens", 8192
+        )
+
+        # Estimate tokens for prompt, input, and output
+        prompt_tokens = count_tokens(op_config["prompt"], model)
+        sample_input = input_data[:100]
+        sample_output = self._run_operation(op_config, input_data[:100])
+
+        prompt_vars = extract_jinja_variables(op_config["prompt"])
+        prompt_vars = [var.split(".")[-1] for var in prompt_vars]
+        avg_input_tokens = mean(
+            [
+                count_tokens(
+                    json.dumps({k: item[k] for k in prompt_vars if k in item}), model
+                )
+                for item in sample_input
+            ]
+        )
+        avg_output_tokens = mean(
+            [
+                count_tokens(
+                    json.dumps({k: item[k] for k in prompt_vars if k in item}), model
+                )
+                for item in sample_output
+            ]
+        )
+
+        # Calculate max batch size that fits in context window
+        max_batch_size = (
+            model_input_context_length - prompt_tokens - avg_output_tokens
+        ) // avg_input_tokens
+
+        # Generate 6 candidate batch sizes
+        batch_sizes = [
+            max(1, int(max_batch_size * ratio))
+            for ratio in [0.1, 0.2, 0.4, 0.6, 0.75, 0.9]
+        ]
+        # Log the generated batch sizes
+        self.console.log("[cyan]Generating plans for batch sizes:[/cyan]")
+        for size in batch_sizes:
+            self.console.log(f"  - {size}")
+        batch_sizes = sorted(set(batch_sizes))  # Remove duplicates and sort
+
+        plans = []
+
+        # Generate multiple fold prompts
+        max_retries = 5
+        retry_count = 0
+        fold_prompts = []
+
+        while retry_count < max_retries and not fold_prompts:
+            try:
+                fold_prompts = self._synthesize_fold_prompts(
+                    op_config,
+                    sample_input,
+                    sample_output,
+                    num_prompts=self.num_fold_prompts,
+                )
+                fold_prompts = list(set(fold_prompts))
+                if not fold_prompts:
+                    raise ValueError("No fold prompts generated")
+            except Exception as e:
+                retry_count += 1
+                if retry_count == max_retries:
+                    raise RuntimeError(
+                        f"Failed to generate fold prompts after {max_retries} attempts: {str(e)}"
+                    )
+                self.console.log(
+                    f"Retry {retry_count}/{max_retries}: Failed to generate fold prompts. Retrying..."
+                )
+
+        for batch_size in batch_sizes:
+            for fold_idx, fold_prompt in enumerate(fold_prompts):
+                plan = op_config.copy()
+                plan["fold_prompt"] = fold_prompt
+                plan["fold_batch_size"] = batch_size
+                plan["associative"] = is_associative
+                plan["name"] = f"{op_config['name']}_bs_{batch_size}_fp_{fold_idx}"
+                plans.append(plan)
+
+        return plans
+
+    def _calculate_compression_ratio(
+        self,
+        op_config: Dict[str, Any],
+        sample_input: List[Dict[str, Any]],
+        sample_output: List[Dict[str, Any]],
+    ) -> float:
+        """
+        Calculate the compression ratio of the reduce operation.
+
+        This method compares the size of the input data to the size of the output data
+        to determine how much the data is being compressed by the reduce operation.
+
+        Args:
+            op_config (Dict[str, Any]): Configuration for the reduce operation.
+            sample_input (List[Dict[str, Any]]): Sample input data.
+            sample_output (List[Dict[str, Any]]): Sample output data.
+
+        Returns:
+            float: The calculated compression ratio.
+        """
+        reduce_key = op_config["reduce_key"]
+        input_schema = op_config.get("input", {}).get("schema", {})
+        output_schema = op_config["output"]["schema"]
+        model = op_config.get("model", "gpt-4o-mini")
+
+        compression_ratios = {}
+
+        # Handle both single key and list of keys
+        if isinstance(reduce_key, list):
+            distinct_keys = set(
+                tuple(item[k] for k in reduce_key) for item in sample_input
+            )
+        else:
+            distinct_keys = set(item[reduce_key] for item in sample_input)
+
+        for key in distinct_keys:
+            if isinstance(reduce_key, list):
+                key_input = [
+                    item
+                    for item in sample_input
+                    if tuple(item[k] for k in reduce_key) == key
+                ]
+                key_output = [
+                    item
+                    for item in sample_output
+                    if tuple(item[k] for k in reduce_key) == key
+                ]
+            else:
+                key_input = [item for item in sample_input if item[reduce_key] == key]
+                key_output = [item for item in sample_output if item[reduce_key] == key]
+
+            if input_schema:
+                key_input_tokens = sum(
+                    count_tokens(
+                        json.dumps({k: item[k] for k in input_schema if k in item}),
+                        model,
+                    )
+                    for item in key_input
+                )
+            else:
+                key_input_tokens = sum(
+                    count_tokens(json.dumps(item), model) for item in key_input
+                )
+
+            key_output_tokens = sum(
+                count_tokens(
+                    json.dumps({k: item[k] for k in output_schema if k in item}), model
+                )
+                for item in key_output
+            )
+
+            compression_ratios[key] = (
+                key_output_tokens / key_input_tokens if key_input_tokens > 0 else 1
+            )
+
+        if not compression_ratios:
+            return 1
+
+        # Calculate importance weights based on the number of items for each key
+        total_items = len(sample_input)
+        if isinstance(reduce_key, list):
+            importance_weights = {
+                key: len(
+                    [
+                        item
+                        for item in sample_input
+                        if tuple(item[k] for k in reduce_key) == key
+                    ]
+                )
+                / total_items
+                for key in compression_ratios
+            }
+        else:
+            importance_weights = {
+                key: len([item for item in sample_input if item[reduce_key] == key])
+                / total_items
+                for key in compression_ratios
+            }
+
+        # Calculate weighted average of compression ratios
+        weighted_sum = sum(
+            compression_ratios[key] * importance_weights[key]
+            for key in compression_ratios
+        )
+        return weighted_sum
+
+    def _synthesize_fold_prompts(
+        self,
+        op_config: Dict[str, Any],
+        sample_input: List[Dict[str, Any]],
+        sample_output: List[Dict[str, Any]],
+        num_prompts: int = 2,
+    ) -> List[str]:
+        """
+        Synthesize fold prompts for the reduce operation. We generate multiple
+        fold prompts in case one is bad.
+
+        A fold operation is a higher-order function that iterates through a data structure,
+        accumulating the results of applying a given combining operation to its elements.
+        In the context of reduce operations, folding allows processing of data in batches,
+        which can significantly improve performance for large datasets.
+
+        This method generates multiple fold prompts that can be used to optimize the reduce operation
+        by allowing it to run on batches of inputs. It uses the language model to create prompts
+        that are variations of the original reduce prompt, adapted for folding operations.
+
+        Args:
+            op_config (Dict[str, Any]): The configuration of the reduce operation.
+            sample_input (List[Dict[str, Any]]): A sample of the input data.
+            sample_output (List[Dict[str, Any]]): A sample of the output data.
+            num_prompts (int, optional): The number of fold prompts to generate. Defaults to 2.
+
+        Returns:
+            List[str]: A list of synthesized fold prompts.
+
+        The method performs the following steps:
+        1. Sets up the system prompt and parameters for the language model.
+        2. Defines a function to get random examples from the sample data.
+        3. Creates a prompt template for generating fold prompts.
+        4. Uses multi-threading to generate multiple fold prompts in parallel.
+        5. Returns the list of generated fold prompts.
+        """
+        system_prompt = "You are an AI assistant tasked with creating a fold prompt for reduce operations in data processing pipelines."
+        original_prompt = op_config["prompt"]
+
+        input_schema = op_config.get("input", {}).get("schema", {})
+        output_schema = op_config["output"]["schema"]
+
+        def get_random_examples():
+            reduce_key = op_config["reduce_key"]
+            reduce_key = list(reduce_key) if not isinstance(reduce_key, list) else reduce_key
+
+            if reduce_key == ["_all"]:
+                # For _all case, just pick random input and output examples
+                input_example = random.choice(sample_input)
+                output_example = random.choice(sample_output)
+            elif isinstance(reduce_key, list):
+                random_key = tuple(
+                    random.choice(
+                        [
+                            tuple(item[k] for k in reduce_key if k in item)
+                            for item in sample_input
+                            if all(k in item for k in reduce_key)
+                        ]
+                    )
+                )
+                input_example = random.choice(
+                    [
+                        item
+                        for item in sample_input
+                        if all(item.get(k) == v for k, v in zip(reduce_key, random_key))
+                    ]
+                )
+                output_example = random.choice(
+                    [
+                        item
+                        for item in sample_output
+                        if all(item.get(k) == v for k, v in zip(reduce_key, random_key))
+                    ]
+                )
+
+            if input_schema:
+                input_example = {
+                    k: input_example[k] for k in input_schema if k in input_example
+                }
+            output_example = {
+                k: output_example[k] for k in output_schema if k in output_example
+            }
+            return input_example, output_example
+
+        parameters = {
+            "type": "object",
+            "properties": {
+                "fold_prompt": {
+                    "type": "string",
+                }
+            },
+            "required": ["fold_prompt"],
+        }
+
+        def generate_single_prompt():
+            input_example, output_example = get_random_examples()
+            prompt = f"""
+            Original Reduce Operation Prompt:
+            {original_prompt}
+
+            Sample Input:
+            {json.dumps(input_example, indent=2)}
+
+            Sample Output:
+            {json.dumps(output_example, indent=2)}
+
+            Create a fold prompt for the reduce operation to run on batches of inputs. The fold prompt should:
+            1. Minimally modify the original reduce prompt
+            2. Describe how to combine the new values with the current reduced value
+            3. Be designed to work iteratively, allowing for multiple fold operations. The first iteration will use the original prompt, and all successive iterations will use the fold prompt.
+
+            The fold prompt should be a Jinja2 template with the following variables available:
+            - {{{{ output }}}}: The current reduced value (a dictionary with the current output schema)
+            - {{{{ inputs }}}}: A list of new values to be folded in
+            - {{{{ reduce_key }}}}: The key used for grouping in the reduce operation
+
+            Provide the fold prompt as a string.
+            """
+            response = self.llm_client.generate(
+                [{"role": "user", "content": prompt}],
+                system_prompt,
+                parameters,
+            )
+            fold_prompt = json.loads(response.choices[0].message.content)["fold_prompt"]
+
+            # Run the operation with the fold prompt
+            # Create a temporary plan with the fold prompt
+            temp_plan = op_config.copy()
+            temp_plan["fold_prompt"] = fold_prompt
+            temp_plan["fold_batch_size"] = min(
+                len(sample_input), 2
+            )  # Use a small batch size for testing
+
+            # Run the operation with the fold prompt
+            try:
+                self._run_operation(temp_plan, sample_input[: temp_plan["fold_batch_size"]])
+
+                return fold_prompt
+            except Exception as e:
+                self.console.log(f"[red]Error in agent-generated fold prompt: {e}[/red]")
+
+                # Create a default fold prompt that instructs folding new data into existing output
+                fold_prompt = f"""Analyze this batch of data using the following instructions:
+
+{original_prompt}
+
+However, instead of starting fresh, fold your analysis into the existing output that has already been generated. The existing output is provided in the 'output' variable below:
+
+{{{{ output }}}} 
+
+Remember, you must fold the new data into the existing output, do not start fresh."""
+                return fold_prompt
+
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            fold_prompts = list(
+                executor.map(lambda _: generate_single_prompt(), range(num_prompts))
+            )
+
+        return fold_prompts
+
+    def _evaluate_reduce_plans(
+        self,
+        op_config: Dict[str, Any],
+        plans: List[Dict[str, Any]],
+        input_data: List[Dict[str, Any]],
+        validator_prompt: str,
+    ) -> Dict[str, Any]:
+        """
+        Evaluate multiple reduce plans and select the best one.
+
+        This method takes a list of reduce plans, evaluates each one using the input data
+        and a validator prompt, and selects the best plan based on the evaluation scores.
+        It also attempts to create and evaluate a merged plan that enhances the runtime performance
+        of the best plan.
+
+        A merged plan is an optimization technique applied to the best-performing plan
+        that uses the fold operation. It allows the best plan to run even faster by
+        executing parallel folds and then merging the results of these individual folds
+        together. We default to a merge batch size of 2, but one can increase this.
+
+        Args:
+            op_config (Dict[str, Any]): The configuration of the reduce operation.
+            plans (List[Dict[str, Any]]): A list of reduce plans to evaluate.
+            input_data (List[Dict[str, Any]]): The input data to use for evaluation.
+            validator_prompt (str): The prompt to use for validating the output of each plan.
+
+        Returns:
+            Dict[str, Any]: The best reduce plan, either the top-performing original plan
+                            or a merged plan if it performs well enough.
+
+        The method performs the following steps:
+        1. Evaluates each plan using multi-threading.
+        2. Sorts the plans based on their evaluation scores.
+        3. Selects the best plan and attempts to create a merged plan.
+        4. Evaluates the merged plan and compares it to the best original plan.
+        5. Returns either the merged plan or the best original plan based on their scores.
+        """
+        self.console.log("\n[bold]Evaluating Reduce Plans:[/bold]")
+        for i, plan in enumerate(plans):
+            self.console.log(f"Plan {i+1} (batch size: {plan['fold_batch_size']})")
+
+        plan_scores = []
+        plan_outputs = {}
+
+        # Create a fixed random sample for evaluation
+        sample_size = min(100, len(input_data))
+        evaluation_sample = random.sample(input_data, sample_size)
+
+        # Create a fixed set of validation samples
+        validation_inputs = self._create_validation_inputs(
+            evaluation_sample, plan["reduce_key"]
+        )
+
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            futures = [
+                executor.submit(
+                    self._evaluate_single_plan,
+                    plan,
+                    evaluation_sample,
+                    validator_prompt,
+                    validation_inputs,
+                )
+                for plan in plans
+            ]
+            for future in as_completed(futures):
+                plan, score, output = future.result()
+                plan_scores.append((plan, score))
+                plan_outputs[id(plan)] = output
+
+        # Sort plans by score in descending order, then by fold_batch_size in descending order
+        sorted_plans = sorted(
+            plan_scores, key=lambda x: (x[1], x[0]["fold_batch_size"]), reverse=True
+        )
+
+        self.console.log("\n[bold]Reduce Plan Scores:[/bold]")
+        for i, (plan, score) in enumerate(sorted_plans):
+            self.console.log(
+                f"Plan {i+1} (batch size: {plan['fold_batch_size']}): {score:.2f}"
+            )
+
+        best_plan, best_score = sorted_plans[0]
+        self.console.log(
+            f"\n[green]Selected best plan with score: {best_score:.2f} and batch size: {best_plan['fold_batch_size']}[/green]"
+        )
+
+        if op_config.get("synthesize_merge", False):
+            # Create a new plan with merge prompt and updated parameters
+            merged_plan = best_plan.copy()
+
+            # Synthesize merge prompt if it doesn't exist
+            if "merge_prompt" not in merged_plan:
+                merged_plan["merge_prompt"] = self._synthesize_merge_prompt(
+                    merged_plan, plan_outputs[id(best_plan)]
+                )
+                # Print the synthesized merge prompt
+                self.console.log("\n[bold]Synthesized Merge Prompt:[/bold]")
+                self.console.log(merged_plan["merge_prompt"])
+
+            # Set merge_batch_size to 2 and num_parallel_folds to 5
+            merged_plan["merge_batch_size"] = 2
+
+            # Evaluate the merged plan
+            _, merged_plan_score, _, operation_instance = self._evaluate_single_plan(
+                merged_plan,
+                evaluation_sample,
+                validator_prompt,
+                validation_inputs,
+                return_instance=True,
+            )
+
+            # Get the merge and fold times from the operation instance
+            merge_times = operation_instance.merge_times
+            fold_times = operation_instance.fold_times
+            merge_avg_time = mean(merge_times) if merge_times else None
+            fold_avg_time = mean(fold_times) if fold_times else None
+
+            self.console.log("\n[bold]Scores:[/bold]")
+            self.console.log(f"Original plan: {best_score:.2f}")
+            self.console.log(f"Merged plan: {merged_plan_score:.2f}")
+
+            # Compare scores and decide which plan to use
+            if merged_plan_score >= best_score * 0.75:
+                self.console.log(
+                    f"\n[green]Using merged plan with score: {merged_plan_score:.2f}[/green]"
+                )
+                if merge_avg_time and fold_avg_time:
+                    merged_plan["merge_time"] = merge_avg_time
+                    merged_plan["fold_time"] = fold_avg_time
+                return merged_plan
+            else:
+                self.console.log(
+                    f"\n[yellow]Merged plan quality too low. Using original plan with score: {best_score:.2f}[/yellow]"
+                )
+                return best_plan
+        else:
+            return best_plan
+
+    def _evaluate_single_plan(
+        self,
+        plan: Dict[str, Any],
+        input_data: List[Dict[str, Any]],
+        validator_prompt: str,
+        validation_inputs: List[Dict[str, Any]],
+        return_instance: bool = False,
+    ) -> Union[
+        Tuple[Dict[str, Any], float, List[Dict[str, Any]]],
+        Tuple[Dict[str, Any], float, List[Dict[str, Any]], BaseOperation],
+    ]:
+        """
+        Evaluate a single reduce plan using the provided input data and validator prompt.
+
+        This method runs the reduce operation with the given plan, validates the output,
+        and calculates a score based on the validation results. The scoring works as follows:
+        1. It counts the number of valid results from the validation.
+        2. The score is calculated as the ratio of valid results to the total number of validation results.
+        3. This produces a score between 0 and 1, where 1 indicates all results were valid, and 0 indicates none were valid.
+
+        TODO: We should come up with a better scoring method here, maybe pairwise comparisons.
+
+        Args:
+            plan (Dict[str, Any]): The reduce plan to evaluate.
+            input_data (List[Dict[str, Any]]): The input data to use for evaluation.
+            validator_prompt (str): The prompt to use for validating the output.
+            return_instance (bool, optional): Whether to return the operation instance. Defaults to False.
+
+        Returns:
+            Union[
+                Tuple[Dict[str, Any], float, List[Dict[str, Any]]],
+                Tuple[Dict[str, Any], float, List[Dict[str, Any]], BaseOperation],
+            ]: A tuple containing the plan, its score, the output data, and optionally the operation instance.
+
+        The method performs the following steps:
+        1. Runs the reduce operation with the given plan on the input data.
+        2. Validates the output using the validator prompt.
+        3. Calculates a score based on the validation results.
+        4. Returns the plan, score, output data, and optionally the operation instance.
+        """
+        output = self._run_operation(plan, input_data, return_instance)
+        if return_instance:
+            output, operation_instance = output
+
+        validation_result = self._validate_reduce_output(
+            plan, validation_inputs, output, validator_prompt
+        )
+
+        # Calculate a score based on validation results
+        valid_count = sum(
+            1
+            for result in validation_result["validation_results"]
+            if result["is_correct"]
+        )
+        score = valid_count / len(validation_result["validation_results"])
+
+        if return_instance:
+            return plan, score, output, operation_instance
+        else:
+            return plan, score, output
+
+    def _synthesize_merge_prompt(
+        self, plan: Dict[str, Any], sample_outputs: List[Dict[str, Any]]
+    ) -> str:
+        """
+        Synthesize a merge prompt for combining multiple folded outputs in a reduce operation.
+
+        This method generates a merge prompt that can be used to combine the results of multiple
+        parallel fold operations into a single output. It uses the language model to create a prompt
+        that is consistent with the original reduce and fold prompts while addressing the specific
+        requirements of merging multiple outputs.
+
+        Args:
+            plan (Dict[str, Any]): The reduce plan containing the original prompt and fold prompt.
+            sample_outputs (List[Dict[str, Any]]): Sample outputs from the fold operation to use as examples.
+
+        Returns:
+            str: The synthesized merge prompt as a string.
+
+        The method performs the following steps:
+        1. Sets up the system prompt for the language model.
+        2. Prepares a random sample output to use as an example.
+        3. Creates a detailed prompt for the language model, including the original reduce prompt,
+           fold prompt, sample output, and instructions for creating the merge prompt.
+        4. Uses the language model to generate the merge prompt.
+        5. Returns the generated merge prompt.
+        """
+        system_prompt = "You are an AI assistant tasked with creating a merge prompt for reduce operations in data processing pipelines. The pipeline has a reduce operation, and incrementally folds inputs into a single output. We want to optimize the pipeline for speed by running multiple folds on different inputs in parallel, and then merging the fold outputs into a single output."
+
+        output_schema = plan["output"]["schema"]
+        random_output = random.choice(sample_outputs)
+        random_output = {
+            k: random_output[k] for k in output_schema if k in random_output
+        }
+
+        prompt = f"""Reduce Operation Prompt (runs on the first batch of inputs):
+        {plan["prompt"]}
+
+        Fold Prompt (runs on the second and subsequent batches of inputs):
+        {plan["fold_prompt"]}
+
+        Sample output of the fold operation (an input to the merge operation):
+        {json.dumps(random_output, indent=2)}
+
+        Create a merge prompt for the reduce operation to combine 2+ folded outputs. The merge prompt should:
+        1. Give context on the task & fold operations, describing that the prompt will be used to combine multiple outputs from the fold operation (as if the original prompt was run on all inputs at once)
+        2. Describe how to combine multiple folded outputs into a single output
+        3. Minimally deviate from the reduce and fold prompts
+
+        The merge prompt should be a Jinja2 template with the following variables available:
+        - {{ outputs }}: A list of reduced outputs to be merged (each following the output schema). You can access the first output with {{ outputs[0] }} and the second with {{ outputs[1] }}
+
+        Output Schema:
+        {json.dumps(output_schema, indent=2)}
+
+        Provide the merge prompt as a string.
+        """
+
+        parameters = {
+            "type": "object",
+            "properties": {
+                "merge_prompt": {
+                    "type": "string",
+                }
+            },
+            "required": ["merge_prompt"],
+        }
+
+        response = self.llm_client.generate(
+            [{"role": "user", "content": prompt}],
+            system_prompt,
+            parameters,
+        )
+        return json.loads(response.choices[0].message.content)["merge_prompt"]
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ __init__(runner, config, console, llm_client, max_threads, run_operation, num_fold_prompts=1, num_samples_in_validation=10, status=None) + +

+ + +
+ +

Initialize the ReduceOptimizer.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ config + + Dict[str, Any] + +
+

Configuration dictionary for the optimizer.

+
+
+ required +
+ console + + Console + +
+

Rich console object for pretty printing.

+
+
+ required +
+ llm_client + + LLMClient + +
+

Client for interacting with a language model.

+
+
+ required +
+ max_threads + + int + +
+

Maximum number of threads to use for parallel processing.

+
+
+ required +
+ run_operation + + Callable + +
+

Function to run an operation.

+
+
+ required +
+ num_fold_prompts + + int + +
+

Number of fold prompts to generate. Defaults to 1.

+
+
+ 1 +
+ num_samples_in_validation + + int + +
+

Number of samples to use in validation. Defaults to 10.

+
+
+ 10 +
+ +
+ Source code in docetl/optimizers/reduce_optimizer.py +
38
+39
+40
+41
+42
+43
+44
+45
+46
+47
+48
+49
+50
+51
+52
+53
+54
+55
+56
+57
+58
+59
+60
+61
+62
+63
+64
+65
+66
+67
+68
+69
+70
def __init__(
+    self,
+    runner,
+    config: Dict[str, Any],
+    console: Console,
+    llm_client: LLMClient,
+    max_threads: int,
+    run_operation: Callable,
+    num_fold_prompts: int = 1,
+    num_samples_in_validation: int = 10,
+    status: Optional[Status] = None,
+):
+    """
+    Initialize the ReduceOptimizer.
+
+    Args:
+        config (Dict[str, Any]): Configuration dictionary for the optimizer.
+        console (Console): Rich console object for pretty printing.
+        llm_client (LLMClient): Client for interacting with a language model.
+        max_threads (int): Maximum number of threads to use for parallel processing.
+        run_operation (Callable): Function to run an operation.
+        num_fold_prompts (int, optional): Number of fold prompts to generate. Defaults to 1.
+        num_samples_in_validation (int, optional): Number of samples to use in validation. Defaults to 10.
+    """
+    self.runner = runner
+    self.config = config
+    self.console = console
+    self.llm_client = llm_client
+    self._run_operation = run_operation
+    self.max_threads = max_threads
+    self.num_fold_prompts = num_fold_prompts
+    self.num_samples_in_validation = num_samples_in_validation
+    self.status = status
+
+
+
+ +
+ +
+ + +

+ optimize(op_config, input_data, level=1) + +

+ + +
+ +

Optimize the reduce operation based on the given configuration and input data.

+

This method performs the following steps: +1. Run the original operation +2. Generate a validator prompt +3. Validate the output +4. If improvement is needed: + a. Evaluate if decomposition is beneficial + b. If decomposition is beneficial, recursively optimize each sub-operation + c. If not, proceed with single operation optimization +5. Run the optimized operation(s)

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ op_config + + Dict[str, Any] + +
+

Configuration for the reduce operation.

+
+
+ required +
+ input_data + + List[Dict[str, Any]] + +
+

Input data for the reduce operation.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + + + + + +
TypeDescription
+ List[Dict[str, Any]] + +
+

Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]: A tuple containing the list of optimized configurations

+
+
+ List[Dict[str, Any]] + +
+

and the list of outputs from the optimized operation(s), and the cost of the operation due to synthesizing any resolve operations.

+
+
+ +
+ Source code in docetl/optimizers/reduce_optimizer.py +
142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
def optimize(
+    self,
+    op_config: Dict[str, Any],
+    input_data: List[Dict[str, Any]],
+    level: int = 1,
+) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]:
+    """
+    Optimize the reduce operation based on the given configuration and input data.
+
+    This method performs the following steps:
+    1. Run the original operation
+    2. Generate a validator prompt
+    3. Validate the output
+    4. If improvement is needed:
+       a. Evaluate if decomposition is beneficial
+       b. If decomposition is beneficial, recursively optimize each sub-operation
+       c. If not, proceed with single operation optimization
+    5. Run the optimized operation(s)
+
+    Args:
+        op_config (Dict[str, Any]): Configuration for the reduce operation.
+        input_data (List[Dict[str, Any]]): Input data for the reduce operation.
+
+    Returns:
+        Tuple[List[Dict[str, Any]], List[Dict[str, Any]], float]: A tuple containing the list of optimized configurations
+        and the list of outputs from the optimized operation(s), and the cost of the operation due to synthesizing any resolve operations.
+    """
+    validation_results, prompt_tokens, model_input_context_length, model, validator_prompt, original_output = self.should_optimize_helper(op_config, input_data)
+
+    add_map_op = False
+    if prompt_tokens * 2 > model_input_context_length:
+        add_map_op = True
+        self.console.log(
+            f"[yellow]Warning: The reduce prompt exceeds the token limit for model {model}. "
+            f"Token count: {prompt_tokens}, Limit: {model_input_context_length}. "
+            f"Add a map operation to the pipeline.[/yellow]"
+        )
+
+    # # Also query an agent to look at a sample of the inputs and see if they think a map operation would be helpful
+    # preprocessing_steps = ""
+    # should_use_map, preprocessing_steps = self._should_use_map(
+    #     op_config, input_data
+    # )
+    # if should_use_map or add_map_op:
+    #     # Synthesize a map operation
+    #     map_prompt, map_output_schema = self._synthesize_map_operation(
+    #         op_config, preprocessing_steps, input_data
+    #     )
+    #     # Change the reduce operation prompt to use the map schema
+    #     new_reduce_prompt = self._change_reduce_prompt_to_use_map_schema(
+    #         op_config["prompt"], map_output_schema
+    #     )
+    #     op_config["prompt"] = new_reduce_prompt
+
+    #     # Return unoptimized map and reduce operations
+    #     return [map_prompt, op_config], input_data, 0.0
+
+
+    # Print the validation results
+    self.console.log("[bold]Validation Results on Initial Sample:[/bold]")
+    if validation_results["needs_improvement"] or self.config.get("optimizer_config", {}).get("force_decompose", False):
+        self.console.post_optimizer_rationale(
+            should_optimize=True,
+            rationale= "\n".join(
+                [
+                    f"Issues: {result['issues']} Suggestions: {result['suggestions']}"
+                    for result in validation_results["validation_results"]
+                ]
+            ),
+            validator_prompt=validator_prompt,
+        )
+        self.console.log(
+            "\n".join(
+                [
+                    f"Issues: {result['issues']} Suggestions: {result['suggestions']}"
+                    for result in validation_results["validation_results"]
+                ]
+            )
+        )
+
+        # Step 3: Evaluate if decomposition is beneficial
+        decomposition_result = self._evaluate_decomposition(
+            op_config, input_data, level
+        )
+
+        if decomposition_result["should_decompose"]:
+            return self._optimize_decomposed_reduce(
+                decomposition_result, op_config, input_data, level
+            )
+
+        return self._optimize_single_reduce(op_config, input_data, validator_prompt)
+    else:
+        self.console.log(f"No improvements identified; {validation_results}.")
+        self.console.post_optimizer_rationale(
+            should_optimize=False,
+            rationale="No improvements identified; no optimization recommended.",
+            validator_prompt=validator_prompt,
+        )
+        return [op_config], original_output, 0.0
+
+
+
+ +
+ + + +
+ +
+ +
+ +
+ + + +

+ docetl.optimizers.join_optimizer.JoinOptimizer + + +

+ + +
+ + + + + + + +
+ Source code in docetl/optimizers/join_optimizer.py +
  17
+  18
+  19
+  20
+  21
+  22
+  23
+  24
+  25
+  26
+  27
+  28
+  29
+  30
+  31
+  32
+  33
+  34
+  35
+  36
+  37
+  38
+  39
+  40
+  41
+  42
+  43
+  44
+  45
+  46
+  47
+  48
+  49
+  50
+  51
+  52
+  53
+  54
+  55
+  56
+  57
+  58
+  59
+  60
+  61
+  62
+  63
+  64
+  65
+  66
+  67
+  68
+  69
+  70
+  71
+  72
+  73
+  74
+  75
+  76
+  77
+  78
+  79
+  80
+  81
+  82
+  83
+  84
+  85
+  86
+  87
+  88
+  89
+  90
+  91
+  92
+  93
+  94
+  95
+  96
+  97
+  98
+  99
+ 100
+ 101
+ 102
+ 103
+ 104
+ 105
+ 106
+ 107
+ 108
+ 109
+ 110
+ 111
+ 112
+ 113
+ 114
+ 115
+ 116
+ 117
+ 118
+ 119
+ 120
+ 121
+ 122
+ 123
+ 124
+ 125
+ 126
+ 127
+ 128
+ 129
+ 130
+ 131
+ 132
+ 133
+ 134
+ 135
+ 136
+ 137
+ 138
+ 139
+ 140
+ 141
+ 142
+ 143
+ 144
+ 145
+ 146
+ 147
+ 148
+ 149
+ 150
+ 151
+ 152
+ 153
+ 154
+ 155
+ 156
+ 157
+ 158
+ 159
+ 160
+ 161
+ 162
+ 163
+ 164
+ 165
+ 166
+ 167
+ 168
+ 169
+ 170
+ 171
+ 172
+ 173
+ 174
+ 175
+ 176
+ 177
+ 178
+ 179
+ 180
+ 181
+ 182
+ 183
+ 184
+ 185
+ 186
+ 187
+ 188
+ 189
+ 190
+ 191
+ 192
+ 193
+ 194
+ 195
+ 196
+ 197
+ 198
+ 199
+ 200
+ 201
+ 202
+ 203
+ 204
+ 205
+ 206
+ 207
+ 208
+ 209
+ 210
+ 211
+ 212
+ 213
+ 214
+ 215
+ 216
+ 217
+ 218
+ 219
+ 220
+ 221
+ 222
+ 223
+ 224
+ 225
+ 226
+ 227
+ 228
+ 229
+ 230
+ 231
+ 232
+ 233
+ 234
+ 235
+ 236
+ 237
+ 238
+ 239
+ 240
+ 241
+ 242
+ 243
+ 244
+ 245
+ 246
+ 247
+ 248
+ 249
+ 250
+ 251
+ 252
+ 253
+ 254
+ 255
+ 256
+ 257
+ 258
+ 259
+ 260
+ 261
+ 262
+ 263
+ 264
+ 265
+ 266
+ 267
+ 268
+ 269
+ 270
+ 271
+ 272
+ 273
+ 274
+ 275
+ 276
+ 277
+ 278
+ 279
+ 280
+ 281
+ 282
+ 283
+ 284
+ 285
+ 286
+ 287
+ 288
+ 289
+ 290
+ 291
+ 292
+ 293
+ 294
+ 295
+ 296
+ 297
+ 298
+ 299
+ 300
+ 301
+ 302
+ 303
+ 304
+ 305
+ 306
+ 307
+ 308
+ 309
+ 310
+ 311
+ 312
+ 313
+ 314
+ 315
+ 316
+ 317
+ 318
+ 319
+ 320
+ 321
+ 322
+ 323
+ 324
+ 325
+ 326
+ 327
+ 328
+ 329
+ 330
+ 331
+ 332
+ 333
+ 334
+ 335
+ 336
+ 337
+ 338
+ 339
+ 340
+ 341
+ 342
+ 343
+ 344
+ 345
+ 346
+ 347
+ 348
+ 349
+ 350
+ 351
+ 352
+ 353
+ 354
+ 355
+ 356
+ 357
+ 358
+ 359
+ 360
+ 361
+ 362
+ 363
+ 364
+ 365
+ 366
+ 367
+ 368
+ 369
+ 370
+ 371
+ 372
+ 373
+ 374
+ 375
+ 376
+ 377
+ 378
+ 379
+ 380
+ 381
+ 382
+ 383
+ 384
+ 385
+ 386
+ 387
+ 388
+ 389
+ 390
+ 391
+ 392
+ 393
+ 394
+ 395
+ 396
+ 397
+ 398
+ 399
+ 400
+ 401
+ 402
+ 403
+ 404
+ 405
+ 406
+ 407
+ 408
+ 409
+ 410
+ 411
+ 412
+ 413
+ 414
+ 415
+ 416
+ 417
+ 418
+ 419
+ 420
+ 421
+ 422
+ 423
+ 424
+ 425
+ 426
+ 427
+ 428
+ 429
+ 430
+ 431
+ 432
+ 433
+ 434
+ 435
+ 436
+ 437
+ 438
+ 439
+ 440
+ 441
+ 442
+ 443
+ 444
+ 445
+ 446
+ 447
+ 448
+ 449
+ 450
+ 451
+ 452
+ 453
+ 454
+ 455
+ 456
+ 457
+ 458
+ 459
+ 460
+ 461
+ 462
+ 463
+ 464
+ 465
+ 466
+ 467
+ 468
+ 469
+ 470
+ 471
+ 472
+ 473
+ 474
+ 475
+ 476
+ 477
+ 478
+ 479
+ 480
+ 481
+ 482
+ 483
+ 484
+ 485
+ 486
+ 487
+ 488
+ 489
+ 490
+ 491
+ 492
+ 493
+ 494
+ 495
+ 496
+ 497
+ 498
+ 499
+ 500
+ 501
+ 502
+ 503
+ 504
+ 505
+ 506
+ 507
+ 508
+ 509
+ 510
+ 511
+ 512
+ 513
+ 514
+ 515
+ 516
+ 517
+ 518
+ 519
+ 520
+ 521
+ 522
+ 523
+ 524
+ 525
+ 526
+ 527
+ 528
+ 529
+ 530
+ 531
+ 532
+ 533
+ 534
+ 535
+ 536
+ 537
+ 538
+ 539
+ 540
+ 541
+ 542
+ 543
+ 544
+ 545
+ 546
+ 547
+ 548
+ 549
+ 550
+ 551
+ 552
+ 553
+ 554
+ 555
+ 556
+ 557
+ 558
+ 559
+ 560
+ 561
+ 562
+ 563
+ 564
+ 565
+ 566
+ 567
+ 568
+ 569
+ 570
+ 571
+ 572
+ 573
+ 574
+ 575
+ 576
+ 577
+ 578
+ 579
+ 580
+ 581
+ 582
+ 583
+ 584
+ 585
+ 586
+ 587
+ 588
+ 589
+ 590
+ 591
+ 592
+ 593
+ 594
+ 595
+ 596
+ 597
+ 598
+ 599
+ 600
+ 601
+ 602
+ 603
+ 604
+ 605
+ 606
+ 607
+ 608
+ 609
+ 610
+ 611
+ 612
+ 613
+ 614
+ 615
+ 616
+ 617
+ 618
+ 619
+ 620
+ 621
+ 622
+ 623
+ 624
+ 625
+ 626
+ 627
+ 628
+ 629
+ 630
+ 631
+ 632
+ 633
+ 634
+ 635
+ 636
+ 637
+ 638
+ 639
+ 640
+ 641
+ 642
+ 643
+ 644
+ 645
+ 646
+ 647
+ 648
+ 649
+ 650
+ 651
+ 652
+ 653
+ 654
+ 655
+ 656
+ 657
+ 658
+ 659
+ 660
+ 661
+ 662
+ 663
+ 664
+ 665
+ 666
+ 667
+ 668
+ 669
+ 670
+ 671
+ 672
+ 673
+ 674
+ 675
+ 676
+ 677
+ 678
+ 679
+ 680
+ 681
+ 682
+ 683
+ 684
+ 685
+ 686
+ 687
+ 688
+ 689
+ 690
+ 691
+ 692
+ 693
+ 694
+ 695
+ 696
+ 697
+ 698
+ 699
+ 700
+ 701
+ 702
+ 703
+ 704
+ 705
+ 706
+ 707
+ 708
+ 709
+ 710
+ 711
+ 712
+ 713
+ 714
+ 715
+ 716
+ 717
+ 718
+ 719
+ 720
+ 721
+ 722
+ 723
+ 724
+ 725
+ 726
+ 727
+ 728
+ 729
+ 730
+ 731
+ 732
+ 733
+ 734
+ 735
+ 736
+ 737
+ 738
+ 739
+ 740
+ 741
+ 742
+ 743
+ 744
+ 745
+ 746
+ 747
+ 748
+ 749
+ 750
+ 751
+ 752
+ 753
+ 754
+ 755
+ 756
+ 757
+ 758
+ 759
+ 760
+ 761
+ 762
+ 763
+ 764
+ 765
+ 766
+ 767
+ 768
+ 769
+ 770
+ 771
+ 772
+ 773
+ 774
+ 775
+ 776
+ 777
+ 778
+ 779
+ 780
+ 781
+ 782
+ 783
+ 784
+ 785
+ 786
+ 787
+ 788
+ 789
+ 790
+ 791
+ 792
+ 793
+ 794
+ 795
+ 796
+ 797
+ 798
+ 799
+ 800
+ 801
+ 802
+ 803
+ 804
+ 805
+ 806
+ 807
+ 808
+ 809
+ 810
+ 811
+ 812
+ 813
+ 814
+ 815
+ 816
+ 817
+ 818
+ 819
+ 820
+ 821
+ 822
+ 823
+ 824
+ 825
+ 826
+ 827
+ 828
+ 829
+ 830
+ 831
+ 832
+ 833
+ 834
+ 835
+ 836
+ 837
+ 838
+ 839
+ 840
+ 841
+ 842
+ 843
+ 844
+ 845
+ 846
+ 847
+ 848
+ 849
+ 850
+ 851
+ 852
+ 853
+ 854
+ 855
+ 856
+ 857
+ 858
+ 859
+ 860
+ 861
+ 862
+ 863
+ 864
+ 865
+ 866
+ 867
+ 868
+ 869
+ 870
+ 871
+ 872
+ 873
+ 874
+ 875
+ 876
+ 877
+ 878
+ 879
+ 880
+ 881
+ 882
+ 883
+ 884
+ 885
+ 886
+ 887
+ 888
+ 889
+ 890
+ 891
+ 892
+ 893
+ 894
+ 895
+ 896
+ 897
+ 898
+ 899
+ 900
+ 901
+ 902
+ 903
+ 904
+ 905
+ 906
+ 907
+ 908
+ 909
+ 910
+ 911
+ 912
+ 913
+ 914
+ 915
+ 916
+ 917
+ 918
+ 919
+ 920
+ 921
+ 922
+ 923
+ 924
+ 925
+ 926
+ 927
+ 928
+ 929
+ 930
+ 931
+ 932
+ 933
+ 934
+ 935
+ 936
+ 937
+ 938
+ 939
+ 940
+ 941
+ 942
+ 943
+ 944
+ 945
+ 946
+ 947
+ 948
+ 949
+ 950
+ 951
+ 952
+ 953
+ 954
+ 955
+ 956
+ 957
+ 958
+ 959
+ 960
+ 961
+ 962
+ 963
+ 964
+ 965
+ 966
+ 967
+ 968
+ 969
+ 970
+ 971
+ 972
+ 973
+ 974
+ 975
+ 976
+ 977
+ 978
+ 979
+ 980
+ 981
+ 982
+ 983
+ 984
+ 985
+ 986
+ 987
+ 988
+ 989
+ 990
+ 991
+ 992
+ 993
+ 994
+ 995
+ 996
+ 997
+ 998
+ 999
+1000
+1001
+1002
+1003
+1004
+1005
+1006
+1007
+1008
+1009
+1010
+1011
+1012
+1013
+1014
+1015
+1016
+1017
+1018
+1019
+1020
+1021
+1022
+1023
+1024
+1025
+1026
+1027
+1028
+1029
+1030
+1031
+1032
+1033
+1034
+1035
+1036
+1037
+1038
+1039
+1040
+1041
+1042
+1043
+1044
+1045
+1046
+1047
+1048
+1049
+1050
+1051
+1052
+1053
+1054
+1055
+1056
+1057
+1058
+1059
+1060
+1061
+1062
+1063
+1064
+1065
+1066
+1067
+1068
+1069
+1070
+1071
+1072
+1073
+1074
+1075
+1076
+1077
+1078
+1079
+1080
+1081
+1082
+1083
+1084
+1085
+1086
+1087
+1088
+1089
+1090
+1091
+1092
+1093
+1094
+1095
+1096
+1097
+1098
+1099
+1100
+1101
+1102
+1103
+1104
+1105
+1106
+1107
+1108
+1109
+1110
+1111
+1112
+1113
+1114
+1115
+1116
+1117
+1118
+1119
+1120
+1121
+1122
+1123
+1124
+1125
+1126
+1127
+1128
+1129
+1130
+1131
+1132
+1133
+1134
+1135
+1136
+1137
+1138
+1139
+1140
+1141
+1142
+1143
+1144
+1145
+1146
+1147
+1148
+1149
+1150
+1151
+1152
+1153
+1154
+1155
+1156
+1157
+1158
+1159
+1160
+1161
+1162
+1163
+1164
+1165
+1166
+1167
+1168
+1169
+1170
+1171
+1172
+1173
+1174
+1175
+1176
+1177
+1178
+1179
+1180
+1181
+1182
+1183
+1184
+1185
+1186
+1187
+1188
+1189
+1190
+1191
+1192
+1193
+1194
+1195
+1196
+1197
+1198
+1199
+1200
+1201
+1202
+1203
+1204
+1205
+1206
+1207
+1208
+1209
+1210
+1211
+1212
+1213
+1214
+1215
+1216
+1217
+1218
+1219
+1220
+1221
+1222
+1223
+1224
+1225
+1226
+1227
+1228
+1229
+1230
+1231
+1232
+1233
+1234
+1235
+1236
+1237
+1238
+1239
+1240
+1241
+1242
+1243
+1244
+1245
+1246
+1247
+1248
+1249
+1250
+1251
+1252
+1253
+1254
+1255
+1256
+1257
+1258
+1259
+1260
+1261
+1262
+1263
+1264
+1265
+1266
+1267
+1268
+1269
+1270
+1271
+1272
+1273
+1274
+1275
+1276
+1277
+1278
+1279
+1280
+1281
+1282
+1283
+1284
+1285
+1286
+1287
+1288
+1289
+1290
+1291
+1292
+1293
+1294
+1295
+1296
+1297
+1298
+1299
+1300
+1301
+1302
+1303
+1304
+1305
+1306
+1307
+1308
+1309
+1310
+1311
+1312
+1313
+1314
+1315
+1316
+1317
+1318
+1319
+1320
+1321
+1322
+1323
+1324
+1325
+1326
+1327
+1328
+1329
+1330
+1331
+1332
+1333
+1334
+1335
+1336
+1337
+1338
+1339
+1340
+1341
+1342
+1343
+1344
+1345
+1346
+1347
+1348
+1349
+1350
+1351
+1352
+1353
+1354
+1355
+1356
+1357
+1358
+1359
+1360
+1361
+1362
+1363
+1364
+1365
+1366
+1367
+1368
+1369
+1370
+1371
+1372
+1373
+1374
+1375
+1376
+1377
+1378
+1379
+1380
+1381
+1382
+1383
+1384
+1385
+1386
+1387
+1388
+1389
+1390
+1391
+1392
+1393
+1394
+1395
+1396
+1397
+1398
+1399
+1400
+1401
+1402
+1403
+1404
+1405
+1406
+1407
+1408
+1409
+1410
+1411
+1412
+1413
+1414
+1415
+1416
+1417
+1418
+1419
+1420
+1421
+1422
+1423
+1424
+1425
+1426
+1427
+1428
+1429
+1430
+1431
+1432
+1433
+1434
+1435
+1436
+1437
+1438
+1439
+1440
+1441
+1442
+1443
+1444
+1445
+1446
+1447
+1448
+1449
+1450
+1451
+1452
+1453
+1454
+1455
+1456
+1457
+1458
+1459
+1460
+1461
+1462
+1463
+1464
+1465
+1466
+1467
+1468
+1469
+1470
+1471
+1472
+1473
+1474
+1475
+1476
+1477
+1478
+1479
+1480
+1481
+1482
+1483
+1484
+1485
+1486
+1487
+1488
+1489
+1490
+1491
+1492
+1493
+1494
+1495
+1496
+1497
+1498
+1499
+1500
+1501
+1502
+1503
+1504
+1505
+1506
+1507
+1508
+1509
+1510
+1511
+1512
+1513
+1514
+1515
+1516
+1517
+1518
+1519
+1520
+1521
+1522
+1523
+1524
+1525
+1526
+1527
+1528
+1529
+1530
+1531
+1532
+1533
+1534
+1535
+1536
+1537
+1538
+1539
+1540
+1541
+1542
+1543
+1544
+1545
+1546
+1547
+1548
+1549
+1550
+1551
+1552
+1553
+1554
+1555
+1556
+1557
+1558
+1559
+1560
+1561
+1562
+1563
+1564
+1565
+1566
+1567
+1568
+1569
+1570
+1571
+1572
+1573
+1574
+1575
+1576
+1577
+1578
+1579
+1580
+1581
+1582
+1583
+1584
+1585
+1586
+1587
+1588
+1589
+1590
+1591
+1592
+1593
+1594
+1595
+1596
+1597
+1598
+1599
+1600
+1601
+1602
+1603
+1604
+1605
+1606
+1607
+1608
+1609
+1610
+1611
+1612
+1613
+1614
+1615
+1616
+1617
+1618
+1619
+1620
+1621
+1622
+1623
+1624
+1625
+1626
+1627
+1628
+1629
+1630
+1631
+1632
+1633
+1634
+1635
+1636
+1637
+1638
+1639
+1640
+1641
+1642
+1643
+1644
+1645
+1646
+1647
+1648
+1649
+1650
+1651
+1652
+1653
+1654
+1655
+1656
+1657
+1658
+1659
+1660
+1661
+1662
+1663
+1664
+1665
+1666
+1667
+1668
+1669
+1670
+1671
+1672
+1673
+1674
+1675
+1676
+1677
+1678
+1679
+1680
+1681
+1682
+1683
+1684
+1685
+1686
+1687
+1688
+1689
+1690
+1691
+1692
+1693
+1694
+1695
+1696
+1697
+1698
+1699
+1700
+1701
+1702
+1703
+1704
+1705
+1706
+1707
+1708
+1709
+1710
+1711
+1712
+1713
+1714
+1715
+1716
+1717
+1718
+1719
+1720
+1721
+1722
+1723
+1724
+1725
class JoinOptimizer:
+    def __init__(
+        self,
+        runner,
+        config: Dict[str, Any],
+        op_config: Dict[str, Any],
+        console: Console,
+        llm_client: Any,
+        max_threads: int,
+        target_recall: float = 0.95,
+        sample_size: int = 500,
+        sampling_weight: float = 20,
+        agent_max_retries: int = 5,
+        estimated_selectivity: float = None,
+        status: Status = None,
+    ):
+        self.runner = runner
+        self.config = config
+        self.op_config = op_config
+        self.llm_client = llm_client
+        self.max_threads = max_threads
+        self.console = console
+        self.target_recall = target_recall
+        self.sample_size = sample_size
+        self.sampling_weight = sampling_weight
+        self.agent_max_retries = agent_max_retries
+        self.estimated_selectivity = estimated_selectivity
+        self.console.log(f"Target Recall: {self.target_recall}")
+        self.status = status
+        # if self.estimated_selectivity is not None:
+        #     self.console.log(
+        #         f"[yellow]Using estimated selectivity of {self.estimated_selectivity}[/yellow]"
+        #     )
+
+    def _analyze_map_prompt_categorization(self, map_prompt: str) -> Tuple[bool, str]:
+        """
+        Analyze the map prompt to determine if it's explicitly categorical.
+
+        Args:
+            map_prompt (str): The map prompt to analyze.
+
+        Returns:
+            bool: True if the prompt is explicitly categorical, False otherwise.
+        """
+        messages = [
+            {
+                "role": "system",
+                "content": "You are an AI assistant tasked with analyzing prompts for data processing operations.",
+            },
+            {
+                "role": "user",
+                "content": f"""Analyze the following map operation prompt and determine if it is explicitly categorical,
+                meaning it details a specific set of possible outputs:
+
+                {map_prompt}
+
+                Respond with 'Yes' if the prompt is explicitly categorical, detailing a finite set of possible outputs.
+                Respond with 'No' if the prompt allows for open-ended or non-categorical responses.
+                Provide a brief explanation for your decision.""",
+            },
+        ]
+
+        response = self.llm_client.generate(
+            messages,
+            "You are an expert in analyzing natural language prompts for data processing tasks.",
+            {
+                "type": "object",
+                "properties": {
+                    "is_categorical": {
+                        "type": "string",
+                        "enum": ["Yes", "No"],
+                        "description": "Whether the prompt is explicitly categorical",
+                    },
+                    "explanation": {
+                        "type": "string",
+                        "description": "Brief explanation for the decision",
+                    },
+                },
+                "required": ["is_categorical", "explanation"],
+            },
+        )
+
+        analysis = json.loads(response.choices[0].message.content)
+
+        self.console.log("[bold]Map Prompt Analysis:[/bold]")
+        self.console.log(f"Is Categorical: {analysis['is_categorical']}")
+        self.console.log(f"Explanation: {analysis['explanation']}")
+
+        return analysis["is_categorical"].lower() == "yes", analysis["explanation"]
+
+    def _determine_duplicate_keys(
+        self,
+        input_data: List[Dict[str, Any]],
+        reduce_key: List[str],
+        map_prompt: Optional[str] = None,
+    ) -> Tuple[bool, str]:
+        # Prepare a sample of the input data for analysis
+        sample_size = min(10, len(input_data))
+        data_sample = random.sample(
+            [{rk: item[rk] for rk in reduce_key} for item in input_data], sample_size
+        )
+
+        context_prefix = ""
+        if map_prompt:
+            context_prefix = f"For context, these values came out of a pipeline with the following prompt:\n\n{map_prompt}\n\n"
+
+        messages = [
+            {
+                "role": "user",
+                "content": f"{context_prefix}I want to do a reduce operation on these values, and I need to determine if there are semantic duplicates in the data, where the strings are different but they technically belong in the same group. Note that exact string duplicates should not be considered here.\n\nHere's a sample of the data (showing the '{reduce_key}' field(s)): {data_sample}\n\nBased on this {'context and ' if map_prompt else ''}sample, are there likely to be such semantic duplicates (not exact string matches) in the dataset? Respond with 'yes' only if you think there are semantic duplicates, or 'no' if you don't see evidence of semantic duplicates or if you only see exact string duplicates.",
+            },
+        ]
+        response = self.llm_client.generate(
+            messages,
+            "You are an expert data analyst. Analyze the given data sample and determine if there are likely to be semantic duplicate values that belong in the same group, even if the strings are different.",
+            {
+                "type": "object",
+                "properties": {
+                    "likely_duplicates": {
+                        "type": "string",
+                        "enum": ["Yes", "No"],
+                        "description": "Whether duplicates are likely to exist in the full dataset",
+                    },
+                    "explanation": {
+                        "type": "string",
+                        "description": "Brief explanation for the decision",
+                    },
+                },
+                "required": ["likely_duplicates", "explanation"],
+            },
+        )
+
+        analysis = json.loads(response.choices[0].message.content)
+
+        self.console.log(f"[bold]Duplicate Analysis for '{reduce_key}':[/bold]")
+        self.console.log(f"Likely Duplicates: {analysis['likely_duplicates']}")
+        self.console.log(f"Explanation: {analysis['explanation']}")
+
+        if analysis["likely_duplicates"].lower() == "yes":
+            self.console.log(
+                "[yellow]Duplicates are likely. Consider using a deduplication strategy in the resolution step.[/yellow]"
+            )
+            return True, analysis["explanation"]
+        return False, ""
+
+    def _sample_random_pairs(
+        self, input_data: List[Dict[str, Any]], n: int
+    ) -> List[Tuple[int, int]]:
+        """Sample random pairs of indices, excluding exact matches."""
+        pairs = set()
+        max_attempts = n * 10  # Avoid infinite loop
+        attempts = 0
+
+        while len(pairs) < n and attempts < max_attempts:
+            i, j = random.sample(range(len(input_data)), 2)
+            if i != j and input_data[i] != input_data[j]:
+                pairs.add((min(i, j), max(i, j)))  # Ensure ordered pairs
+            attempts += 1
+
+        return list(pairs)
+
+    def _check_duplicates_with_llm(
+        self,
+        input_data: List[Dict[str, Any]],
+        pairs: List[Tuple[int, int]],
+        reduce_key: List[str],
+        map_prompt: Optional[str],
+    ) -> Tuple[bool, str]:
+        """Use LLM to check if any pairs are duplicates."""
+
+        content = "Analyze the following pairs of entries and determine if any of them are likely duplicates. Respond with 'Yes' if you find any likely duplicates, or 'No' if none of the pairs seem to be duplicates. Provide a brief explanation for your decision.\n\n"
+
+        if map_prompt:
+            content = (
+                f"For reference, here is the map prompt used earlier in the pipeline: {map_prompt}\n\n"
+                + content
+            )
+
+        for i, (idx1, idx2) in enumerate(pairs, 1):
+            content += f"Pair {i}:\n"
+            content += "Entry 1:\n"
+            for key in reduce_key:
+                content += f"{key}: {json.dumps(input_data[idx1][key], indent=2)}\n"
+            content += "\nEntry 2:\n"
+            for key in reduce_key:
+                content += f"{key}: {json.dumps(input_data[idx2][key], indent=2)}\n"
+            content += "\n"
+
+        messages = [{"role": "user", "content": content}]
+
+        system_prompt = "You are an AI assistant tasked with identifying potential duplicate entries in a dataset."
+        response_schema = {
+            "type": "object",
+            "properties": {
+                "duplicates_found": {"type": "string", "enum": ["Yes", "No"]},
+                "explanation": {"type": "string"},
+            },
+            "required": ["duplicates_found", "explanation"],
+        }
+
+        response = self.llm_client.generate(messages, system_prompt, response_schema)
+
+        # Print the duplicates_found and explanation
+        self.console.log(
+            f"[bold]Duplicates in keys found:[/bold] {response['duplicates_found']}\n"
+            f"[bold]Explanation:[/bold] {response['explanation']}"
+        )
+
+        return response["duplicates_found"].lower() == "yes", response["explanation"]
+
+    def synthesize_compare_prompt(
+        self, map_prompt: Optional[str], reduce_key: List[str]
+    ) -> str:
+
+        system_prompt = f"You are an AI assistant tasked with creating a comparison prompt for LLM-assisted entity resolution. Your task is to create a comparison prompt that will be used to compare two entities, referred to as input1 and input2, to see if they are likely the same entity based on the following reduce key(s): {', '.join(reduce_key)}."
+        if map_prompt:
+            system_prompt += f"\n\nFor context, here is the prompt used earlier in the pipeline to create the inputs to resolve: {map_prompt}"
+
+        messages = [
+            {
+                "role": "user",
+                "content": f"""
+    Create a comparison prompt for entity resolution: The prompt should:
+    1. Be tailored to the specific domain and type of data being compared ({reduce_key}), based on the context provided.
+    2. Instruct to compare two entities, referred to as input1 and input2.
+    3. Specifically mention comparing each reduce key in input1 and input2 (e.g., input1.{{key}} and input2.{{key}} for each key in {reduce_key}). You can reference other fields in the input as well, as long as they are short.
+    4. Include instructions to consider relevant attributes or characteristics for comparison.
+    5. Ask to respond with "True" if the entities are likely the same, or "False" if they are likely different.
+
+    Example structure:
+    ```
+    Compare the following two {reduce_key} from [entity or document type]:
+
+    [Entity 1]:
+    {{{{ input1.key1 }}}}
+    {{{{ input1.optional_key2 }}}}
+
+    [Entity 2]:
+    {{{{ input2.key1 }}}}
+    {{{{ input2.optional_key2 }}}}
+
+    Are these [entities] likely referring to the same [entity type]? Consider [list relevant attributes or characteristics to compare]. Respond with "True" if they are likely the same [entity type], or "False" if they are likely different [entity types].
+    ```
+
+    Please generate the comparison prompt, which should be a Jinja2 template:
+    """,
+            }
+        ]
+
+        response = self.llm_client.generate(
+            messages,
+            system_prompt,
+            {
+                "type": "object",
+                "properties": {
+                    "comparison_prompt": {
+                        "type": "string",
+                        "description": "Detailed comparison prompt for entity resolution",
+                    }
+                },
+                "required": ["comparison_prompt"],
+            },
+        )
+
+        comparison_prompt = json.loads(response.choices[0].message.content)[
+            "comparison_prompt"
+        ]
+
+        # Log the synthesized comparison prompt
+        self.console.log("[green]Synthesized comparison prompt:[/green]")
+        self.console.log(comparison_prompt)
+
+        if not comparison_prompt:
+            raise ValueError(
+                "Could not synthesize a comparison prompt. Please provide a comparison prompt in the config."
+            )
+
+        return comparison_prompt
+
+    def synthesize_resolution_prompt(
+        self,
+        map_prompt: Optional[str],
+        reduce_key: List[str],
+        output_schema: Dict[str, str],
+    ) -> str:
+        system_prompt = f"""You are an AI assistant tasked with creating a resolution prompt for LLM-assisted entity resolution.
+        Your task is to create a prompt that will be used to merge multiple duplicate keys into a single, consolidated key.
+        The key(s) being resolved (known as the reduce_key) are {', '.join(reduce_key)}.
+        The duplicate keys will be provided in a list called 'inputs' in a Jinja2 template.
+        """
+
+        if map_prompt:
+            system_prompt += f"\n\nFor context, here is the prompt used earlier in the pipeline to create the inputs to resolve: {map_prompt}"
+
+        messages = [
+            {
+                "role": "user",
+                "content": f"""
+    Create a resolution prompt for merging duplicate keys into a single key. The prompt should:
+    1. Be tailored to the specific domain and type of data being merged, based on the context provided.
+    2. Use a Jinja2 template to iterate over the duplicate keys (accessed as 'inputs', where each item is a dictionary containing the reduce_key fields, which you can access as entry.reduce_key for each reduce_key in {reduce_key}).
+    3. Instruct to create a single, consolidated key from the duplicate keys.
+    4. Include guidelines for resolving conflicts (e.g., choosing the most recent, most complete, or most reliable information).
+    5. Specify that the output of the resolution prompt should conform to the given output schema: {json.dumps(output_schema, indent=2)}
+
+    Example structure:
+    ```
+    Analyze the following duplicate entries for the {reduce_key} key:
+
+    {{% for key in inputs %}}
+    Entry {{{{ loop.index }}}}:
+    {{ % for key in reduce_key %}}
+    {{{{ key }}}}: {{{{ key[reduce_key] }}}}
+    {{% endfor %}}
+
+    {{% endfor %}}
+
+    Merge these into a single key.
+    When merging, follow these guidelines:
+    1. [Provide specific merging instructions relevant to the data type]
+    2. [Do not make the prompt too long]
+
+    Ensure that the merged key conforms to the following schema:
+    {json.dumps(output_schema, indent=2)}
+
+    Return the consolidated key as a single [appropriate data type] value.
+    ```
+
+    Please generate the resolution prompt:
+    """,
+            }
+        ]
+
+        response = self.llm_client.generate(
+            messages,
+            system_prompt,
+            {
+                "type": "object",
+                "properties": {
+                    "resolution_prompt": {
+                        "type": "string",
+                        "description": "Detailed resolution prompt for merging duplicate keys",
+                    }
+                },
+                "required": ["resolution_prompt"],
+            },
+        )
+
+        resolution_prompt = json.loads(response.choices[0].message.content)[
+            "resolution_prompt"
+        ]
+
+        # Log the synthesized resolution prompt
+        self.console.log("[green]Synthesized resolution prompt:[/green]")
+        self.console.log(resolution_prompt)
+
+        if not resolution_prompt:
+            raise ValueError(
+                "Could not synthesize a resolution prompt. Please provide a resolution prompt in the config."
+            )
+
+        return resolution_prompt
+
+    def should_optimize(self, input_data: List[Dict[str, Any]]) -> Tuple[bool, str]:
+        """
+        Determine if the given operation configuration should be optimized.
+        """
+        # If there are no blocking keys or embeddings, then we don't need to optimize
+        if not self.op_config.get("blocking_conditions") or not self.op_config.get("blocking_threshold"):
+            return True, ""
+
+        # Check if the operation is marked as empty
+        elif self.op_config.get("empty", False):
+            # Extract the map prompt from the intermediates
+            map_prompt = self.op_config["_intermediates"]["map_prompt"]
+            reduce_key = self.op_config["_intermediates"]["reduce_key"]
+
+            if reduce_key is None:
+                raise ValueError(
+                    "[yellow]Warning: No reduce key found in intermediates for synthesized resolve operation.[/yellow]"
+                )
+
+            dedup = True
+            explanation = "There is a reduce operation that does not follow a resolve operation. Consider adding a resolve operation to deduplicate the data."
+
+            if map_prompt:
+                # Analyze the map prompt
+                analysis, explanation = self._analyze_map_prompt_categorization(map_prompt)
+
+                if analysis:
+                    dedup = False
+            else:
+                self.console.log(
+                    "[yellow]No map prompt found in intermediates for analysis.[/yellow]"
+                )
+
+            # TODO: figure out why this would ever be the case
+            if not map_prompt:
+                map_prompt = "N/A"
+
+            if dedup is False:
+                dedup, explanation = self._determine_duplicate_keys(
+                    input_data, reduce_key, map_prompt
+                )
+
+            # Now do the last attempt of pairwise comparisons
+            if dedup is False:
+                # Sample up to 20 random pairs of keys for duplicate analysis
+                sampled_pairs = self._sample_random_pairs(input_data, 20)
+
+                # Use LLM to check for duplicates
+                duplicates_found, explanation = self._check_duplicates_with_llm(
+                    input_data, sampled_pairs, reduce_key, map_prompt
+                )
+
+                if duplicates_found:
+                    dedup = True
+
+            return dedup, explanation
+
+        return False, ""
+
+    def optimize_resolve(
+        self, input_data: List[Dict[str, Any]]
+    ) -> Tuple[Dict[str, Any], float]:
+
+        # Check if the operation is marked as empty
+        if self.op_config.get("empty", False):
+            # Extract the map prompt from the intermediates
+            dedup, _ = self.should_optimize(input_data)
+            reduce_key = self.op_config["_intermediates"]["reduce_key"]
+            map_prompt = self.op_config["_intermediates"]["map_prompt"]
+
+            if dedup is False:
+                # If no deduplication is needed, return the same config with 0 cost
+                return self.op_config, 0.0
+
+            # Add the reduce key to the output schema in the config
+            self.op_config["output"] = {"schema": {rk: "string" for rk in reduce_key}}
+            for attempt in range(2):  # Try up to 2 times
+                self.op_config["comparison_prompt"] = self.synthesize_compare_prompt(
+                    map_prompt, reduce_key
+                )
+                if (
+                    "input1" in self.op_config["comparison_prompt"]
+                    and "input2" in self.op_config["comparison_prompt"]
+                ):
+                    break
+                elif attempt == 0:
+                    self.console.log(
+                        "[yellow]Warning: 'input1' or 'input2' not found in comparison prompt. Retrying...[/yellow]"
+                    )
+            if (
+                "input1" not in self.op_config["comparison_prompt"]
+                or "input2" not in self.op_config["comparison_prompt"]
+            ):
+                self.console.log(
+                    "[red]Error: Failed to generate comparison prompt with 'input1' and 'input2'. Using last generated prompt.[/red]"
+                )
+            for attempt in range(2):  # Try up to 2 times
+                self.op_config["resolution_prompt"] = self.synthesize_resolution_prompt(
+                    map_prompt, reduce_key, self.op_config["output"]["schema"]
+                )
+                if "inputs" in self.op_config["resolution_prompt"]:
+                    break
+                elif attempt == 0:
+                    self.console.log(
+                        "[yellow]Warning: 'inputs' not found in resolution prompt. Retrying...[/yellow]"
+                    )
+            if "inputs" not in self.op_config["resolution_prompt"]:
+                self.console.log(
+                    "[red]Error: Failed to generate resolution prompt with 'inputs'. Using last generated prompt.[/red]"
+                )
+
+            # Pop off the empty flag
+            self.op_config.pop("empty")
+
+        embeddings, blocking_keys, embedding_cost = self._compute_embeddings(input_data)
+        self.console.log(
+            f"[bold]Cost of creating embeddings on the sample: ${embedding_cost:.4f}[/bold]"
+        )
+
+        similarities = self._calculate_cosine_similarities(embeddings)
+
+        sampled_pairs = self._sample_pairs(similarities)
+        comparison_results, comparison_cost = self._perform_comparisons_resolve(
+            input_data, sampled_pairs
+        )
+
+        self._print_similarity_histogram(similarities, comparison_results)
+
+        threshold, estimated_selectivity = self._find_optimal_threshold(
+            comparison_results, similarities
+        )
+
+        blocking_rules = self._generate_blocking_rules(
+            blocking_keys, input_data, comparison_results
+        )
+
+        if blocking_rules:
+            false_negatives, rule_selectivity = self._verify_blocking_rule(
+                input_data,
+                blocking_rules[0],
+                blocking_keys,
+                comparison_results,
+            )
+            if not false_negatives and rule_selectivity <= estimated_selectivity:
+                self.console.log(
+                    "[green]Blocking rule verified. No false negatives detected in the sample and selectivity is within estimated selectivity.[/green]"
+                )
+            else:
+                if false_negatives:
+                    self.console.log(
+                        f"[red]Blocking rule rejected. {len(false_negatives)} false negatives detected in the sample.[/red]"
+                    )
+                    for i, j in false_negatives[:5]:  # Show up to 5 examples
+                        self.console.log(
+                            f"  Filtered pair: {{ {blocking_keys[0]}: {input_data[i][blocking_keys[0]]} }} and {{ {blocking_keys[0]}: {input_data[j][blocking_keys[0]]} }}"
+                        )
+                    if len(false_negatives) > 5:
+                        self.console.log(f"  ... and {len(false_negatives) - 5} more.")
+                if rule_selectivity > estimated_selectivity:
+                    self.console.log(
+                        f"[red]Blocking rule rejected. Rule selectivity ({rule_selectivity:.4f}) is higher than the estimated selectivity ({estimated_selectivity:.4f}).[/red]"
+                    )
+                blocking_rules = (
+                    []
+                )  # Clear the blocking rule if it introduces false negatives or is too selective
+
+        optimized_config = self._update_config(threshold, blocking_keys, blocking_rules)
+        return optimized_config, embedding_cost + comparison_cost
+
+    def optimize_equijoin(
+        self, left_data: List[Dict[str, Any]], right_data: List[Dict[str, Any]]
+    ) -> Tuple[Dict[str, Any], float, Dict[str, Any]]:
+        left_keys = self.op_config.get("blocking_keys", {}).get("left", [])
+        right_keys = self.op_config.get("blocking_keys", {}).get("right", [])
+
+        if not left_keys and not right_keys:
+            # Ask the LLM agent if it would be beneficial to do a map operation on
+            # one of the datasets before doing an equijoin
+            apply_transformation, dataset_to_transform, reason = (
+                self._should_apply_map_transformation(
+                    left_keys, right_keys, left_data, right_data
+                )
+            )
+
+            if apply_transformation:
+                self.console.log(
+                    f"LLM agent suggested applying a map transformation to {dataset_to_transform} dataset because: {reason}"
+                )
+                extraction_prompt, output_key, new_comparison_prompt = (
+                    self._generate_map_and_new_join_transformation(
+                        dataset_to_transform, reason, left_data, right_data
+                    )
+                )
+                self.console.log(
+                    f"Generated map transformation prompt: {extraction_prompt}"
+                )
+                self.console.log(f"\nNew output key: {output_key}")
+                self.console.log(
+                    f"\nNew equijoin comparison prompt: {new_comparison_prompt}"
+                )
+
+                # Update the comparison prompt
+                self.op_config["comparison_prompt"] = new_comparison_prompt
+
+                # Add the output key to the left_keys or right_keys
+                if dataset_to_transform == "left":
+                    left_keys.append(output_key)
+                else:
+                    right_keys.append(output_key)
+
+                # Reset the blocking keys in the config
+                self.op_config["blocking_keys"] = {
+                    "left": left_keys,
+                    "right": right_keys,
+                }
+
+                # Bubble up this config and return the transformation prompt, so we can optimize the map operation
+                return (
+                    self.op_config,
+                    0.0,
+                    {
+                        "optimize_map": True,
+                        "map_prompt": extraction_prompt,
+                        "output_key": output_key,
+                        "dataset_to_transform": dataset_to_transform,
+                    },
+                )
+
+            # Print the reason for not applying a map transformation
+            self.console.log(
+                f"Reason for not synthesizing a map transformation for either left or right dataset: {reason}"
+            )
+
+        # If there are no blocking keys, generate them
+        if not left_keys or not right_keys:
+            generated_left_keys, generated_right_keys = (
+                self._generate_blocking_keys_equijoin(left_data, right_data)
+            )
+            left_keys.extend(generated_left_keys)
+            right_keys.extend(generated_right_keys)
+            left_keys = list(set(left_keys))
+            right_keys = list(set(right_keys))
+
+            # Log the generated blocking keys
+            self.console.log(
+                f"[bold]Generated blocking keys (for embeddings-based blocking):[/bold]"
+            )
+            self.console.log(f"Left keys: {left_keys}")
+            self.console.log(f"Right keys: {right_keys}")
+
+        left_embeddings, _, left_embedding_cost = self._compute_embeddings(
+            left_data, keys=left_keys
+        )
+        right_embeddings, _, right_embedding_cost = self._compute_embeddings(
+            right_data, keys=right_keys
+        )
+        self.console.log(
+            f"[bold]Cost of creating embeddings on the sample: ${left_embedding_cost + right_embedding_cost:.4f}[/bold]"
+        )
+
+        similarities = self._calculate_cross_similarities(
+            left_embeddings, right_embeddings
+        )
+
+        sampled_pairs = self._sample_pairs(similarities)
+        comparison_results, comparison_cost = self._perform_comparisons_equijoin(
+            left_data, right_data, sampled_pairs
+        )
+        self._print_similarity_histogram(similarities, comparison_results)
+        while not any(result[2] for result in comparison_results):
+            self.console.log(
+                "[yellow]No matches found in the current sample. Resampling pairs to compare...[/yellow]"
+            )
+            sampled_pairs = self._sample_pairs(similarities)
+            comparison_results, current_cost = self._perform_comparisons_equijoin(
+                left_data, right_data, sampled_pairs
+            )
+            comparison_cost += current_cost
+            self._print_similarity_histogram(similarities, comparison_results)
+
+        threshold, estimated_selectivity = self._find_optimal_threshold(
+            comparison_results, similarities
+        )
+        self.estimated_selectivity = estimated_selectivity
+
+        blocking_rules = self._generate_blocking_rules_equijoin(
+            left_keys, right_keys, left_data, right_data, comparison_results
+        )
+
+        if blocking_rules:
+            false_negatives, rule_selectivity = self._verify_blocking_rule_equijoin(
+                left_data,
+                right_data,
+                blocking_rules[0],
+                left_keys,
+                right_keys,
+                comparison_results,
+            )
+            if not false_negatives and rule_selectivity <= estimated_selectivity:
+                self.console.log(
+                    "[green]Blocking rule verified. No false negatives detected in the sample and selectivity is within bounds.[/green]"
+                )
+            else:
+                if false_negatives:
+                    self.console.log(
+                        f"[red]Blocking rule rejected. {len(false_negatives)} false negatives detected in the sample.[/red]"
+                    )
+                    for i, j in false_negatives[:5]:  # Show up to 5 examples
+                        self.console.log(
+                            f"  Filtered pair: Left: {{{', '.join(f'{key}: {left_data[i][key]}' for key in left_keys)}}} and Right: {{{', '.join(f'{key}: {right_data[j][key]}' for key in right_keys)}}}"
+                        )
+                    if len(false_negatives) > 5:
+                        self.console.log(f"  ... and {len(false_negatives) - 5} more.")
+                if rule_selectivity > estimated_selectivity:
+                    self.console.log(
+                        f"[red]Blocking rule rejected. Rule selectivity ({rule_selectivity:.4f}) is higher than the estimated selectivity ({estimated_selectivity:.4f}).[/red]"
+                    )
+                blocking_rules = (
+                    []
+                )  # Clear the blocking rule if it introduces false negatives or is too selective
+
+        containment_rules = self._generate_containment_rules_equijoin(
+            left_data, right_data
+        )
+        self.console.log(
+            f"[bold]Generated {len(containment_rules)} containment rules. Please select which ones to use as blocking conditions:[/bold]"
+        )
+        selected_containment_rules = []
+        for rule in containment_rules:
+            self.console.log(f"[green]{rule}[/green]")
+            # Temporarily stop the status
+            if self.status:
+                self.status.stop()
+            # Use Rich's Confirm for input
+            if Confirm.ask("Use this rule?", self.console):
+                selected_containment_rules.append(rule)
+            # Restart the status
+            if self.status:
+                self.status.start()
+
+        if len(containment_rules) > 0:
+            self.console.log(
+                f"[bold]Selected {len(selected_containment_rules)} containment rules for blocking.[/bold]"
+            )
+        blocking_rules.extend(selected_containment_rules)
+
+        optimized_config = self._update_config_equijoin(
+            threshold, left_keys, right_keys, blocking_rules
+        )
+        return (
+            optimized_config,
+            left_embedding_cost + right_embedding_cost + comparison_cost,
+            {},
+        )
+
+    def _should_apply_map_transformation(
+        self,
+        left_keys: List[str],
+        right_keys: List[str],
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+        sample_size: int = 5,
+    ) -> Tuple[bool, str, str]:
+        # Sample data
+        left_sample = random.sample(left_data, min(sample_size, len(left_data)))
+        right_sample = random.sample(right_data, min(sample_size, len(right_data)))
+
+        # Get keys and their average lengths
+        all_left_keys = {
+            k: sum(len(str(d[k])) for d in left_sample) / len(left_sample)
+            for k in left_sample[0].keys()
+        }
+        all_right_keys = {
+            k: sum(len(str(d[k])) for d in right_sample) / len(right_sample)
+            for k in right_sample[0].keys()
+        }
+
+        messages = [
+            {
+                "role": "user",
+                "content": f"""Analyze the following datasets and determine if an additional LLM transformation should be applied to generate a new key-value pair for easier joining:
+
+                Comparison prompt for the join operation: {self.op_config.get('comparison_prompt', 'No comparison prompt provided.')}
+
+                Left dataset keys and average lengths: {json.dumps(all_left_keys, indent=2)}
+                Right dataset keys and average lengths: {json.dumps(all_right_keys, indent=2)}
+
+                Left dataset sample:
+                {json.dumps(left_sample, indent=2)}
+
+                Right dataset sample:
+                {json.dumps(right_sample, indent=2)}
+
+                Current keys used for embedding-based ranking of likely matches:
+                Left keys: {left_keys}
+                Right keys: {right_keys}
+
+                Consider the following:
+                1. Are the current keys sufficient for accurate embedding-based ranking of likely matches? We don't want to use too many keys, or keys with too much information, as this will dilute the signal in the embeddings.
+                2. Are there any keys particularly long (e.g., full text fields), containing information that is not relevant for the join operation?
+                3. Is there information spread across multiple fields that could be combined?
+                4. Would a summary or extraction of key information be beneficial?
+                5. Is there a mismatch in information representation between the datasets?
+                6. Could an additional LLM-generated field improve the accuracy of embeddings or join comparisons?
+
+                If you believe an additional LLM transformation would be beneficial, specify which dataset (left or right) should be transformed and explain why. In most cases, you should pick the dataset with the longer keys unless there is a specific reason to pick the other dataset. Otherwise, indicate that no additional transformation is needed and explain why the current blocking keys are sufficient.""",
+            }
+        ]
+
+        response = self.llm_client.generate(
+            messages,
+            "You are an AI expert in data analysis and entity matching.",
+            {
+                "type": "object",
+                "properties": {
+                    "apply_transformation": {"type": "boolean"},
+                    "dataset_to_transform": {
+                        "type": "string",
+                        "enum": ["left", "right", "none"],
+                    },
+                    "reason": {"type": "string"},
+                },
+                "required": ["apply_transformation", "dataset_to_transform", "reason"],
+            },
+        )
+
+        result = json.loads(response.choices[0].message.content)
+
+        return (
+            result["apply_transformation"],
+            result["dataset_to_transform"],
+            result["reason"],
+        )
+
+    def _generate_map_and_new_join_transformation(
+        self,
+        dataset_to_transform: str,
+        reason: str,
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+        sample_size: int = 5,
+    ) -> Tuple[str, str, str]:
+        # Sample data
+        left_sample = random.sample(left_data, min(sample_size, len(left_data)))
+        right_sample = random.sample(right_data, min(sample_size, len(right_data)))
+
+        target_data = left_sample if dataset_to_transform == "left" else right_sample
+
+        messages = [
+            {
+                "role": "user",
+                "content": f"""Generate an LLM prompt to transform the {dataset_to_transform} dataset for easier joining. The transformation should create a new key-value pair.
+
+                Current comparison prompt for the join operation: {self.op_config.get('comparison_prompt', 'No comparison prompt provided.')}
+
+                Target ({dataset_to_transform}) dataset sample:
+                {json.dumps(target_data, indent=2)}
+
+                Other ({'left' if dataset_to_transform == "right" else "right"}) dataset sample:
+                {json.dumps(right_sample if dataset_to_transform == "left" else left_sample, indent=2)}
+
+                Reason for transforming {dataset_to_transform} dataset: {reason}
+
+                Please provide:
+                1. An LLM prompt to extract a smaller representation of what is relevant to the join task. The prompt should be a Jinja2 template, referring to any fields in the input data as {{ input.field_name }}. The prompt should instruct the LLM to return some **non-empty** string-valued output. The transformation should be tailored to the join task if possible, not just a generic summary of the data.
+                2. A name for the new output key that will store the transformed data.
+                3. An edited comparison prompt that leverages the new attribute created by the transformation. This prompt should be a Jinja2 template, referring to any fields in the input data as {{ left.field_name }} and {{ right.field_name }}. The prompt should be the same as the current comparison prompt, but with a new instruction that leverages the new attribute created by the transformation. The prompt should instruct the LLM to return a boolean-valued output, like the current comparison prompt.""",
+            }
+        ]
+
+        response = self.llm_client.generate(
+            messages,
+            "You are an AI expert in data analysis and decomposing complex data processing pipelines.",
+            {
+                "type": "object",
+                "properties": {
+                    "extraction_prompt": {"type": "string"},
+                    "output_key": {"type": "string"},
+                    "new_comparison_prompt": {"type": "string"},
+                },
+                "required": [
+                    "extraction_prompt",
+                    "output_key",
+                    "new_comparison_prompt",
+                ],
+            },
+        )
+
+        result = json.loads(response.choices[0].message.content)
+
+        return (
+            result["extraction_prompt"],
+            result["output_key"],
+            result["new_comparison_prompt"],
+        )
+
+    def _generate_blocking_keys_equijoin(
+        self,
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+        sample_size: int = 5,
+    ) -> Tuple[List[str], List[str]]:
+        # Sample data
+        left_sample = random.sample(left_data, min(sample_size, len(left_data)))
+        right_sample = random.sample(right_data, min(sample_size, len(right_data)))
+
+        # Prepare sample data for LLM
+        left_keys = list(left_sample[0].keys())
+        right_keys = list(right_sample[0].keys())
+
+        messages = [
+            {
+                "role": "user",
+                "content": f"""Given the following sample data from two datasets, select appropriate blocking keys for an equijoin operation.
+                The blocking process works as follows:
+                1. We create embeddings for the selected keys from both datasets.
+                2. We use cosine similarity between these embeddings to filter pairs for more detailed LLM comparison.
+                3. Pairs with high similarity will be passed to the LLM for final comparison.
+
+                The blocking keys should have relatively short values and be useful for generating embeddings that capture the essence of potential matches.
+
+                Left dataset keys: {left_keys}
+                Right dataset keys: {right_keys}
+
+                Sample from left dataset:
+                {json.dumps(left_sample, indent=2)}
+
+                Sample from right dataset:
+                {json.dumps(right_sample, indent=2)}
+
+                For context, here is the comparison prompt that will be used for the more detailed LLM comparison:
+                {self.op_config.get('comparison_prompt', 'No comparison prompt provided.')}
+
+                Please select one or more keys from each dataset that would be suitable for blocking. The keys should contain information that's likely to be similar in matching records and align with the comparison prompt's focus.""",
+            }
+        ]
+
+        response = self.llm_client.generate(
+            messages,
+            "You are an expert in entity matching and database operations.",
+            {
+                "type": "object",
+                "properties": {
+                    "left_blocking_keys": {
+                        "type": "array",
+                        "items": {"type": "string"},
+                        "description": "List of selected blocking keys from the left dataset",
+                    },
+                    "right_blocking_keys": {
+                        "type": "array",
+                        "items": {"type": "string"},
+                        "description": "List of selected blocking keys from the right dataset",
+                    },
+                },
+                "required": ["left_blocking_keys", "right_blocking_keys"],
+            },
+        )
+
+        result = json.loads(response.choices[0].message.content)
+        left_blocking_keys = result["left_blocking_keys"]
+        right_blocking_keys = result["right_blocking_keys"]
+
+        return left_blocking_keys, right_blocking_keys
+
+    def _compute_embeddings(
+        self,
+        input_data: List[Dict[str, Any]],
+        keys: List[str] = None,
+        is_join: bool = True,
+    ) -> Tuple[List[List[float]], List[str], float]:
+        if keys is None:
+            keys = self.op_config.get("blocking_keys", [])
+            if not keys:
+                prompt_template = self.op_config.get("comparison_prompt", "")
+                prompt_vars = extract_jinja_variables(prompt_template)
+                # Get rid of input, input1, input2
+                prompt_vars = [
+                    var
+                    for var in prompt_vars
+                    if var not in ["input", "input1", "input2"]
+                ]
+
+                # strip all things before . in the prompt_vars
+                keys += list(set([var.split(".")[-1] for var in prompt_vars]))
+            if not keys:
+                self.console.log(
+                    "[yellow]Warning: No blocking keys found. Using all keys for blocking.[/yellow]"
+                )
+                keys = list(input_data[0].keys())
+
+        model_input_context_length = model_cost.get(
+            self.op_config.get("embedding_model", "text-embedding-3-small"), {}
+        ).get("max_input_tokens", 8192)
+        texts = [
+            " ".join(str(item[key]) for key in keys if key in item)[
+                :model_input_context_length
+            ]
+            for item in input_data
+        ]
+
+        embeddings = []
+        total_cost = 0
+        batch_size = 2000
+        for i in range(0, len(texts), batch_size):
+            batch = texts[i : i + batch_size]
+            self.console.log(
+                f"[cyan]Processing batch {i//batch_size + 1} of {len(texts)//batch_size + 1}[/cyan]"
+            )
+            response = self.runner.api.gen_embedding(
+                model=self.op_config.get("embedding_model", "text-embedding-3-small"),
+                input=batch,
+            )
+            embeddings.extend([data["embedding"] for data in response["data"]])
+            total_cost += completion_cost(response)
+        embeddings = [data["embedding"] for data in response["data"]]
+        cost = completion_cost(response)
+        return embeddings, keys, cost
+
+    def _calculate_cosine_similarities(
+        self, embeddings: List[List[float]]
+    ) -> List[Tuple[int, int, float]]:
+        embeddings_array = np.array(embeddings)
+        norms = np.linalg.norm(embeddings_array, axis=1)
+        dot_products = np.dot(embeddings_array, embeddings_array.T)
+        similarities_matrix = dot_products / np.outer(norms, norms)
+        i, j = np.triu_indices(len(embeddings), k=1)
+        similarities = list(
+            zip(i.tolist(), j.tolist(), similarities_matrix[i, j].tolist())
+        )
+        return similarities
+
+    def _print_similarity_histogram(
+        self,
+        similarities: List[Tuple[int, int, float]],
+        comparison_results: List[Tuple[int, int, bool]],
+    ):
+        flat_similarities = [sim[-1] for sim in similarities if sim[-1] != 1]
+        hist, bin_edges = np.histogram(flat_similarities, bins=20)
+        max_bar_width, max_count = 50, max(hist)
+        normalized_hist = [int(count / max_count * max_bar_width) for count in hist]
+
+        # Create a dictionary to store true labels
+        true_labels = {(i, j): is_match for i, j, is_match in comparison_results}
+
+        self.console.log("\n[bold]Embedding Cosine Similarity Distribution:[/bold]")
+        for i, count in enumerate(normalized_hist):
+            bar = "█" * count
+            label = f"{bin_edges[i]:.2f}-{bin_edges[i+1]:.2f}"
+
+            # Count true matches and not matches in this bin
+            true_matches = 0
+            not_matches = 0
+            labeled_count = 0
+            for sim in similarities:
+                if bin_edges[i] <= sim[2] < bin_edges[i + 1]:
+                    if (sim[0], sim[1]) in true_labels:
+                        labeled_count += 1
+                        if true_labels[(sim[0], sim[1])]:
+                            true_matches += 1
+                        else:
+                            not_matches += 1
+
+            # Calculate percentages of labeled pairs
+            if labeled_count > 0:
+                true_match_percent = (true_matches / labeled_count) * 100
+                not_match_percent = (not_matches / labeled_count) * 100
+            else:
+                true_match_percent = 0
+                not_match_percent = 0
+
+            self.console.log(
+                f"{label}: {bar} "
+                f"(Labeled: {labeled_count}/{hist[i]}, [green]{true_match_percent:.1f}% match[/green], [red]{not_match_percent:.1f}% not match[/red])"
+            )
+        self.console.log("\n")
+
+    def _sample_pairs(
+        self, similarities: List[Tuple[int, int, float]]
+    ) -> List[Tuple[int, int]]:
+        # Sort similarities in descending order
+        sorted_similarities = sorted(similarities, key=lambda x: x[2], reverse=True)
+
+        # Calculate weights using exponential weighting with self.sampling_weight
+        similarities_array = np.array([sim[2] for sim in sorted_similarities])
+        weights = np.exp(self.sampling_weight * similarities_array)
+        weights /= weights.sum()  # Normalize weights to sum to 1
+
+        # Sample pairs based on the calculated weights
+        sampled_indices = np.random.choice(
+            len(sorted_similarities),
+            size=min(self.sample_size, len(sorted_similarities)),
+            replace=False,
+            p=weights,
+        )
+
+        sampled_pairs = [
+            (sorted_similarities[i][0], sorted_similarities[i][1])
+            for i in sampled_indices
+        ]
+        return sampled_pairs
+
+    def _calculate_cross_similarities(
+        self, left_embeddings: List[List[float]], right_embeddings: List[List[float]]
+    ) -> List[Tuple[int, int, float]]:
+        left_array = np.array(left_embeddings)
+        right_array = np.array(right_embeddings)
+        dot_product = np.dot(left_array, right_array.T)
+        norm_left = np.linalg.norm(left_array, axis=1)
+        norm_right = np.linalg.norm(right_array, axis=1)
+        similarities = dot_product / np.outer(norm_left, norm_right)
+        return [
+            (i, j, sim)
+            for i, row in enumerate(similarities)
+            for j, sim in enumerate(row)
+        ]
+
+    def _perform_comparisons_resolve(
+        self, input_data: List[Dict[str, Any]], pairs: List[Tuple[int, int]]
+    ) -> Tuple[List[Tuple[int, int, bool]], float]:
+        comparisons, total_cost = [], 0
+        op = ResolveOperation(
+            self.runner,
+            self.op_config,
+            self.runner.default_model,
+            self.max_threads,
+            self.console,
+            self.status,
+        )
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            futures = [
+                executor.submit(
+                    op.compare_pair,
+                    self.op_config["comparison_prompt"],
+                    self.op_config.get(
+                        "comparison_model", self.config.get("model", "gpt-4o-mini")
+                    ),
+                    input_data[i],
+                    input_data[j],
+                )
+                for i, j in pairs
+            ]
+            for future, (i, j) in zip(futures, pairs):
+                is_match, cost, _ = future.result()
+                comparisons.append((i, j, is_match))
+                total_cost += cost
+
+        self.console.log(
+            f"[bold]Cost of pairwise comparisons on the sample: ${total_cost:.4f}[/bold]"
+        )
+        return comparisons, total_cost
+
+    def _perform_comparisons_equijoin(
+        self,
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+        pairs: List[Tuple[int, int]],
+    ) -> Tuple[List[Tuple[int, int, bool]], float]:
+        comparisons, total_cost = [], 0
+        op = EquijoinOperation(
+            self.runner,
+            self.op_config,
+            self.runner.default_model,
+            self.max_threads,
+            self.console,
+            self.status,
+        )
+        with ThreadPoolExecutor(max_workers=self.max_threads) as executor:
+            futures = [
+                executor.submit(
+                    op.compare_pair,
+                    self.op_config["comparison_prompt"],
+                    self.op_config.get(
+                        "comparison_model", self.config.get("model", "gpt-4o-mini")
+                    ),
+                    left_data[i],
+                    right_data[j] if right_data else left_data[j],
+                )
+                for i, j in pairs
+            ]
+            for future, (i, j) in zip(futures, pairs):
+                is_match, cost = future.result()
+                comparisons.append((i, j, is_match))
+                total_cost += cost
+
+        self.console.log(
+            f"[bold]Cost of pairwise comparisons on the sample: ${total_cost:.4f}[/bold]"
+        )
+        return comparisons, total_cost
+
+    def _find_optimal_threshold(
+        self,
+        comparisons: List[Tuple[int, int, bool]],
+        similarities: List[Tuple[int, int, float]],
+    ) -> Tuple[float, float, float]:
+        true_labels = np.array([comp[2] for comp in comparisons])
+        sim_dict = {(i, j): sim for i, j, sim in similarities}
+        sim_scores = np.array([sim_dict[(i, j)] for i, j, _ in comparisons])
+
+        thresholds = np.linspace(0, 1, 100)
+        precisions, recalls = [], []
+
+        for threshold in thresholds:
+            predictions = sim_scores >= threshold
+            tp = np.sum(predictions & true_labels)
+            fp = np.sum(predictions & ~true_labels)
+            fn = np.sum(~predictions & true_labels)
+
+            precision = tp / (tp + fp) if (tp + fp) > 0 else 0
+            recall = tp / (tp + fn) if (tp + fn) > 0 else 0
+
+            precisions.append(precision)
+            recalls.append(recall)
+
+        valid_indices = [i for i, r in enumerate(recalls) if r >= self.target_recall]
+        if not valid_indices:
+            optimal_threshold = float(thresholds[np.argmax(recalls)])
+        else:
+            optimal_threshold = float(thresholds[max(valid_indices)])
+
+        # Improved selectivity estimation
+        all_similarities = np.array([s[2] for s in similarities])
+        sampled_similarities = sim_scores
+
+        # Calculate sampling probabilities
+        sampling_probs = np.exp(self.sampling_weight * sampled_similarities)
+        sampling_probs /= sampling_probs.sum()
+
+        # Estimate selectivity using importance sampling
+        weights = 1 / (len(all_similarities) * sampling_probs)
+        numerator = np.sum(weights * true_labels)
+        denominator = np.sum(weights)
+        selectivity_estimate = numerator / denominator
+
+        self.console.log(
+            "[bold cyan]┌─ Estimated Self-Join Selectivity ─────────────────────────┐[/bold cyan]"
+        )
+        self.console.log(
+            f"[bold cyan]│[/bold cyan] [yellow]Target Recall:[/yellow] {self.target_recall:.0%}"
+        )
+        self.console.log(
+            f"[bold cyan]│[/bold cyan] [yellow]Estimate:[/yellow] {selectivity_estimate:.4f}"
+        )
+        self.console.log(
+            "[bold cyan]└───────────────────────────────────────────────────────────┘[/bold cyan]"
+        )
+        self.console.log(
+            f"[bold]Chosen similarity threshold for blocking: {optimal_threshold:.4f}[/bold]"
+        )
+
+        return round(optimal_threshold, 4), selectivity_estimate
+
+    def _generate_blocking_rules(
+        self,
+        blocking_keys: List[str],
+        input_data: List[Dict[str, Any]],
+        comparisons: List[Tuple[int, int, bool]],
+    ) -> List[str]:
+        # Sample 2 true and 2 false comparisons
+        true_comparisons = [comp for comp in comparisons if comp[2]][:2]
+        false_comparisons = [comp for comp in comparisons if not comp[2]][:2]
+        sample_datas = [
+            (
+                {key: input_data[i][key] for key in blocking_keys},
+                {key: input_data[j][key] for key in blocking_keys},
+                is_match,
+            )
+            for i, j, is_match in true_comparisons + false_comparisons
+        ]
+
+        messages = [
+            {
+                "role": "user",
+                "content": f"""Given the following sample comparisons between entities, generate a single-line Python statement that acts as a blocking rule for entity resolution. This rule will be used in the form: `eval(blocking_rule, {{"input1": item1, "input2": item2}})`.
+
+    Sample comparisons (note: these are just a few examples and may not represent all possible cases):
+    {json.dumps(sample_datas, indent=2)}
+
+    For context, here is the comparison prompt that will be used for the more expensive, detailed comparison:
+    {self.op_config.get('comparison_prompt', 'No comparison prompt provided.')}
+
+    Please generate ONE one-line blocking rule that adheres to the following criteria:
+    1. The rule should evaluate to True if the entities are possibly a match and require further comparison.
+    2. The rule should evaluate to False ONLY if the entities are definitely not a match.
+    3. The rule must be a single Python expression that can be evaluated using the eval() function.
+    4. The rule should be much faster to evaluate than the full comparison prompt.
+    5. The rule should capture the essence of the comparison prompt but in a simplified manner.
+    6. The rule should be general enough to work well on the entire dataset, not just these specific examples.
+    7. The rule should handle inconsistent casing by using string methods like .lower() when comparing string values.
+    8. The rule should err on the side of inclusivity - it's better to have false positives than false negatives.
+
+    Example structure of a one-line blocking rule:
+    "(condition1) or (condition2) or (condition3)"
+
+    Where conditions could be comparisons like:
+    "input1['field'].lower() == input2['field'].lower()"
+    "abs(len(input1['text']) - len(input2['text'])) <= 5"
+    "any(word in input1['description'].lower() for word in input2['description'].lower().split())"
+
+    If there's no clear rule that can be generated based on the given information, return the string "True" to ensure all pairs are compared.
+
+    Remember, the primary goal of the blocking rule is to safely reduce the number of comparisons by quickly identifying pairs that are definitely not matches, while keeping all potential matches for further evaluation.""",
+            }
+        ]
+
+        for attempt in range(self.agent_max_retries):  # Up to 3 attempts
+            # Generate blocking rule using the LLM
+            response = self.llm_client.generate(
+                messages,
+                "You are an expert in entity resolution and Python programming. Your task is to generate one efficient blocking rule based on the given sample comparisons and data structure.",
+                {
+                    "type": "object",
+                    "properties": {
+                        "blocking_rule": {
+                            "type": "string",
+                            "description": "One-line Python statement acting as a blocking rule",
+                        }
+                    },
+                    "required": ["blocking_rule"],
+                },
+            )
+
+            # Extract the blocking rule from the LLM response
+            blocking_rule = response.choices[0].message.content
+            blocking_rule = json.loads(blocking_rule).get("blocking_rule")
+
+            if blocking_rule:
+                self.console.log("")  # Print a newline
+
+                if blocking_rule.strip() == "True":
+                    self.console.log(
+                        "[yellow]No suitable blocking rule could be found. Proceeding without a blocking rule.[/yellow]"
+                    )
+                    return []
+
+                self.console.log(
+                    f"[bold]Generated blocking rule (Attempt {attempt + 1}):[/bold] {blocking_rule}"
+                )
+
+                # Test the blocking rule
+                filtered_pairs = self._test_blocking_rule(
+                    input_data, blocking_keys, blocking_rule, comparisons
+                )
+
+                if not filtered_pairs:
+                    self.console.log(
+                        "[green]Blocking rule looks good! No known matches were filtered out.[/green]"
+                    )
+                    return [blocking_rule]
+                else:
+                    feedback = f"The previous rule incorrectly filtered out {len(filtered_pairs)} known matches. "
+                    feedback += (
+                        "Here are up to 3 examples of incorrectly filtered pairs:\n"
+                    )
+                    for i, j in filtered_pairs[:3]:
+                        feedback += f"Item 1: {json.dumps({key: input_data[i][key] for key in blocking_keys})}\nItem 2: {json.dumps({key: input_data[j][key] for key in blocking_keys})}\n"
+                        feedback += "These pairs are known matches but were filtered out by the rule.\n"
+                    feedback += "Please generate a new rule that doesn't filter out these matches."
+
+                    messages.append({"role": "assistant", "content": blocking_rule})
+                    messages.append({"role": "user", "content": feedback})
+            else:
+                self.console.log("[yellow]No blocking rule generated.[/yellow]")
+                return []
+
+        self.console.log(
+            f"[yellow]Failed to generate a suitable blocking rule after {self.agent_max_retries} attempts. Proceeding without a blocking rule.[/yellow]"
+        )
+        return []
+
+    def _test_blocking_rule(
+        self,
+        input_data: List[Dict[str, Any]],
+        blocking_keys: List[str],
+        blocking_rule: str,
+        comparisons: List[Tuple[int, int, bool]],
+    ) -> List[Tuple[int, int]]:
+        def apply_blocking_rule(item1, item2):
+            try:
+                return eval(blocking_rule, {"input1": item1, "input2": item2})
+            except Exception as e:
+                self.console.log(f"[red]Error applying blocking rule: {e}[/red]")
+                return True  # If there's an error, we default to comparing the pair
+
+        filtered_pairs = []
+
+        for i, j, is_match in comparisons:
+            if is_match:
+                item1 = {
+                    k: input_data[i][k] for k in blocking_keys if k in input_data[i]
+                }
+                item2 = {
+                    k: input_data[j][k] for k in blocking_keys if k in input_data[j]
+                }
+
+                if not apply_blocking_rule(item1, item2):
+                    filtered_pairs.append((i, j))
+
+        if filtered_pairs:
+            self.console.log(
+                f"[yellow italic]LLM Correction: The blocking rule incorrectly filtered out {len(filtered_pairs)} known positive matches.[/yellow italic]"
+            )
+            for i, j in filtered_pairs[:5]:  # Show up to 5 examples
+                self.console.log(
+                    f"  Incorrectly filtered pair 1: {json.dumps({key: input_data[i][key] for key in blocking_keys})}  and pair 2: {json.dumps({key: input_data[j][key] for key in blocking_keys})}"
+                )
+            if len(filtered_pairs) > 5:
+                self.console.log(
+                    f"  ... and {len(filtered_pairs) - 5} more incorrect pairs."
+                )
+
+        return filtered_pairs
+
+    def _generate_containment_rules_equijoin(
+        self,
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+    ) -> List[str]:
+        # Get all available keys from the sample data
+        left_keys = set(left_data[0].keys())
+        right_keys = set(right_data[0].keys())
+
+        # Sample a few records from each dataset
+        sample_left = random.sample(left_data, min(3, len(left_data)))
+        sample_right = random.sample(right_data, min(3, len(right_data)))
+
+        messages = [
+            {
+                "role": "system",
+                "content": "You are an AI assistant tasked with generating containment-based blocking rules for an equijoin operation.",
+            },
+            {
+                "role": "user",
+                "content": f"""Generate multiple one-line Python statements that act as containment-based blocking rules for equijoin. These rules will be used in the form: `eval(blocking_rule, {{"left": item1, "right": item2}})`.
+
+Available keys in left dataset: {', '.join(left_keys)}
+Available keys in right dataset: {', '.join(right_keys)}
+
+Sample data from left dataset:
+{json.dumps(sample_left, indent=2)}
+
+Sample data from right dataset:
+{json.dumps(sample_right, indent=2)}
+
+Comparison prompt used for detailed comparison:
+{self.op_config.get('comparison_prompt', 'No comparison prompt provided.')}
+
+Please generate multiple one-line blocking rules that adhere to the following criteria:
+1. The rules should focus on containment relationships between fields in the left and right datasets. Containment can mean that the left field contains all the words in the right field, or the right field contains all the words in the left field.
+2. Each rule should evaluate to True if there's a potential match based on containment, False otherwise.
+3. Rules must be single Python expressions that can be evaluated using the eval() function.
+4. Rules should handle inconsistent casing by using string methods like .lower() when comparing string values.
+5. Consider the length of the fields when generating rules: for example, if the left field is much longer than the right field, it's more likely to contain all the words in the right field.
+
+Example structures of containment-based blocking rules:
+"all(word in left['{{left_key}}'].lower() for word in right['{{right_key}}'].lower().split())"
+"any(word in right['{{right_key}}'].lower().split() for word in left['{{left_key}}'].lower().split())"
+
+Please provide 3-5 different containment-based blocking rules, based on the keys and sample data provided.""",
+            },
+        ]
+
+        response = self.llm_client.generate(
+            messages,
+            "You are an expert in data matching and Python programming.",
+            {
+                "type": "object",
+                "properties": {
+                    "containment_rules": {
+                        "type": "array",
+                        "items": {"type": "string"},
+                        "description": "List of containment-based blocking rules as Python expressions",
+                    }
+                },
+                "required": ["containment_rules"],
+            },
+        )
+
+        containment_rules = response.choices[0].message.content
+        containment_rules = json.loads(containment_rules).get("containment_rules")
+        return containment_rules
+
+    def _generate_blocking_rules_equijoin(
+        self,
+        left_keys: List[str],
+        right_keys: List[str],
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+        comparisons: List[Tuple[int, int, bool]],
+    ) -> List[str]:
+        if not left_keys or not right_keys:
+            left_keys = list(left_data[0].keys())
+            right_keys = list(right_data[0].keys())
+
+        # Sample 2 true and 2 false comparisons
+        true_comparisons = [comp for comp in comparisons if comp[2]][:2]
+        false_comparisons = [comp for comp in comparisons if not comp[2]][:2]
+        sample_datas = [
+            (
+                {key: left_data[i][key] for key in left_keys if key in left_data[i]},
+                {key: right_data[j][key] for key in right_keys if key in right_data[j]},
+                is_match,
+            )
+            for i, j, is_match in true_comparisons + false_comparisons
+        ]
+
+        messages = [
+            {
+                "role": "user",
+                "content": f"""Given the following sample comparisons between entities, generate a single-line Python statement that acts as a blocking rule for equijoin. This rule will be used in the form: `eval(blocking_rule, {{"left": item1, "right": item2}})`.
+
+    Sample comparisons (note: these are just a few examples and may not represent all possible cases):
+    {json.dumps(sample_datas, indent=2)}
+
+    For context, here is the comparison prompt that will be used for the more expensive, detailed comparison:
+    {self.op_config.get('comparison_prompt', 'No comparison prompt provided.')}
+
+    Please generate ONE one-line blocking rule that adheres to the following criteria:
+    1. The rule should evaluate to True if the entities are possibly a match and require further comparison.
+    2. The rule should evaluate to False ONLY if the entities are definitely not a match.
+    3. The rule must be a single Python expression that can be evaluated using the eval() function.
+    4. The rule should be much faster to evaluate than the full comparison prompt.
+    5. The rule should capture the essence of the comparison prompt but in a simplified manner.
+    6. The rule should be general enough to work well on the entire dataset, not just these specific examples.
+    7. The rule should handle inconsistent casing by using string methods like .lower() when comparing string values.
+    8. The rule should err on the side of inclusivity - it's better to have false positives than false negatives.
+
+    Example structure of a one-line blocking rule:
+    "(condition1) or (condition2) or (condition3)"
+
+    Where conditions could be comparisons like:
+    "left['{left_keys[0]}'].lower() == right['{right_keys[0]}'].lower()"
+    "abs(len(left['{left_keys[0]}']) - len(right['{right_keys[0]}'])) <= 5"
+    "any(word in left['{left_keys[0]}'].lower() for word in right['{right_keys[0]}'].lower().split())"
+
+    If there's no clear rule that can be generated based on the given information, return the string "True" to ensure all pairs are compared.
+
+    Remember, the primary goal of the blocking rule is to safely reduce the number of comparisons by quickly identifying pairs that are definitely not matches, while keeping all potential matches for further evaluation.""",
+            }
+        ]
+
+        for attempt in range(self.agent_max_retries):
+            response = self.llm_client.generate(
+                messages,
+                "You are an expert in entity resolution and Python programming. Your task is to generate one efficient blocking rule based on the given sample comparisons and data structure.",
+                {
+                    "type": "object",
+                    "properties": {
+                        "blocking_rule": {
+                            "type": "string",
+                            "description": "One-line Python statement acting as a blocking rule",
+                        }
+                    },
+                    "required": ["blocking_rule"],
+                },
+            )
+
+            blocking_rule = response.choices[0].message.content
+            blocking_rule = json.loads(blocking_rule).get("blocking_rule")
+
+            if blocking_rule:
+                self.console.log("")
+
+                if blocking_rule.strip() == "True":
+                    self.console.log(
+                        "[yellow]No suitable blocking rule could be found. Proceeding without a blocking rule.[/yellow]"
+                    )
+                    return []
+
+                self.console.log(
+                    f"[bold]Generated blocking rule (Attempt {attempt + 1}):[/bold] {blocking_rule}"
+                )
+
+                # Test the blocking rule
+                filtered_pairs = self._test_blocking_rule_equijoin(
+                    left_data,
+                    right_data,
+                    left_keys,
+                    right_keys,
+                    blocking_rule,
+                    comparisons,
+                )
+
+                if not filtered_pairs:
+                    self.console.log(
+                        "[green]Blocking rule looks good! No known matches were filtered out.[/green]"
+                    )
+                    return [blocking_rule]
+                else:
+                    feedback = f"The previous rule incorrectly filtered out {len(filtered_pairs)} known matches. "
+                    feedback += (
+                        "Here are up to 3 examples of incorrectly filtered pairs:\n"
+                    )
+                    for i, j in filtered_pairs[:3]:
+                        feedback += f"Left: {json.dumps({key: left_data[i][key] for key in left_keys})}\n"
+                        feedback += f"Right: {json.dumps({key: right_data[j][key] for key in right_keys})}\n"
+                        feedback += "These pairs are known matches but were filtered out by the rule.\n"
+                    feedback += "Please generate a new rule that doesn't filter out these matches."
+
+                    messages.append({"role": "assistant", "content": blocking_rule})
+                    messages.append({"role": "user", "content": feedback})
+            else:
+                self.console.log("[yellow]No blocking rule generated.[/yellow]")
+                return []
+
+        self.console.log(
+            f"[yellow]Failed to generate a suitable blocking rule after {self.agent_max_retries} attempts. Proceeding without a blocking rule.[/yellow]"
+        )
+        return []
+
+    def _test_blocking_rule_equijoin(
+        self,
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+        left_keys: List[str],
+        right_keys: List[str],
+        blocking_rule: str,
+        comparisons: List[Tuple[int, int, bool]],
+    ) -> List[Tuple[int, int]]:
+        def apply_blocking_rule(left, right):
+            try:
+                return eval(blocking_rule, {"left": left, "right": right})
+            except Exception as e:
+                self.console.log(f"[red]Error applying blocking rule: {e}[/red]")
+                return True  # If there's an error, we default to comparing the pair
+
+        filtered_pairs = []
+
+        for i, j, is_match in comparisons:
+            if is_match:
+                left = left_data[i]
+                right = right_data[j]
+                if not apply_blocking_rule(left, right):
+                    filtered_pairs.append((i, j))
+
+        if filtered_pairs:
+            self.console.log(
+                f"[yellow italic]LLM Correction: The blocking rule incorrectly filtered out {len(filtered_pairs)} known positive matches.[/yellow italic]"
+            )
+            for i, j in filtered_pairs[:5]:  # Show up to 5 examples
+                left_dict = {key: left_data[i][key] for key in left_keys}
+                right_dict = {key: right_data[j][key] for key in right_keys}
+                self.console.log(
+                    f"  Incorrectly filtered pair - Left: {json.dumps(left_dict)}  Right: {json.dumps(right_dict)}"
+                )
+            if len(filtered_pairs) > 5:
+                self.console.log(
+                    f"  ... and {len(filtered_pairs) - 5} more incorrect pairs."
+                )
+
+        return filtered_pairs
+
+    def _verify_blocking_rule_equijoin(
+        self,
+        left_data: List[Dict[str, Any]],
+        right_data: List[Dict[str, Any]],
+        blocking_rule: str,
+        left_keys: List[str],
+        right_keys: List[str],
+        comparison_results: List[Tuple[int, int, bool]],
+    ) -> Tuple[List[Tuple[int, int]], float]:
+        def apply_blocking_rule(left, right):
+            try:
+                return eval(blocking_rule, {"left": left, "right": right})
+            except Exception as e:
+                self.console.log(f"[red]Error applying blocking rule: {e}[/red]")
+                return True  # If there's an error, we default to comparing the pair
+
+        false_negatives = []
+        total_pairs = 0
+        blocked_pairs = 0
+
+        for i, j, is_match in comparison_results:
+            total_pairs += 1
+            left = left_data[i]
+            right = right_data[j]
+            if apply_blocking_rule(left, right):
+                blocked_pairs += 1
+                if is_match:
+                    false_negatives.append((i, j))
+
+        rule_selectivity = blocked_pairs / total_pairs if total_pairs > 0 else 0
+
+        return false_negatives, rule_selectivity
+
+    def _update_config_equijoin(
+        self,
+        threshold: float,
+        left_keys: List[str],
+        right_keys: List[str],
+        blocking_rules: List[str],
+    ) -> Dict[str, Any]:
+        optimized_config = self.op_config.copy()
+        optimized_config["blocking_keys"] = {
+            "left": left_keys,
+            "right": right_keys,
+        }
+        optimized_config["blocking_threshold"] = threshold
+        if blocking_rules:
+            optimized_config["blocking_conditions"] = blocking_rules
+        if "embedding_model" not in optimized_config:
+            optimized_config["embedding_model"] = "text-embedding-3-small"
+        return optimized_config
+
+    def _verify_blocking_rule(
+        self,
+        input_data: List[Dict[str, Any]],
+        blocking_rule: str,
+        blocking_keys: List[str],
+        comparison_results: List[Tuple[int, int, bool]],
+    ) -> Tuple[List[Tuple[int, int]], float]:
+        def apply_blocking_rule(item1, item2):
+            try:
+                return eval(blocking_rule, {"input1": item1, "input2": item2})
+            except Exception as e:
+                self.console.log(f"[red]Error applying blocking rule: {e}[/red]")
+                return True  # If there's an error, we default to comparing the pair
+
+        false_negatives = []
+        total_pairs = 0
+        blocked_pairs = 0
+
+        for i, j, is_match in comparison_results:
+            total_pairs += 1
+            item1 = {k: input_data[i][k] for k in blocking_keys if k in input_data[i]}
+            item2 = {k: input_data[j][k] for k in blocking_keys if k in input_data[j]}
+
+            if apply_blocking_rule(item1, item2):
+                blocked_pairs += 1
+                if is_match:
+                    false_negatives.append((i, j))
+
+        rule_selectivity = blocked_pairs / total_pairs if total_pairs > 0 else 0
+
+        return false_negatives, rule_selectivity
+
+    def _update_config(
+        self, threshold: float, blocking_keys: List[str], blocking_rules: List[str]
+    ) -> Dict[str, Any]:
+        optimized_config = self.op_config.copy()
+        optimized_config["blocking_keys"] = blocking_keys
+        optimized_config["blocking_threshold"] = threshold
+        if blocking_rules:
+            optimized_config["blocking_conditions"] = blocking_rules
+        if "embedding_model" not in optimized_config:
+            optimized_config["embedding_model"] = "text-embedding-3-small"
+        return optimized_config
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ should_optimize(input_data) + +

+ + +
+ +

Determine if the given operation configuration should be optimized.

+ +
+ Source code in docetl/optimizers/join_optimizer.py +
380
+381
+382
+383
+384
+385
+386
+387
+388
+389
+390
+391
+392
+393
+394
+395
+396
+397
+398
+399
+400
+401
+402
+403
+404
+405
+406
+407
+408
+409
+410
+411
+412
+413
+414
+415
+416
+417
+418
+419
+420
+421
+422
+423
+424
+425
+426
+427
+428
+429
+430
+431
+432
+433
+434
+435
+436
+437
def should_optimize(self, input_data: List[Dict[str, Any]]) -> Tuple[bool, str]:
+    """
+    Determine if the given operation configuration should be optimized.
+    """
+    # If there are no blocking keys or embeddings, then we don't need to optimize
+    if not self.op_config.get("blocking_conditions") or not self.op_config.get("blocking_threshold"):
+        return True, ""
+
+    # Check if the operation is marked as empty
+    elif self.op_config.get("empty", False):
+        # Extract the map prompt from the intermediates
+        map_prompt = self.op_config["_intermediates"]["map_prompt"]
+        reduce_key = self.op_config["_intermediates"]["reduce_key"]
+
+        if reduce_key is None:
+            raise ValueError(
+                "[yellow]Warning: No reduce key found in intermediates for synthesized resolve operation.[/yellow]"
+            )
+
+        dedup = True
+        explanation = "There is a reduce operation that does not follow a resolve operation. Consider adding a resolve operation to deduplicate the data."
+
+        if map_prompt:
+            # Analyze the map prompt
+            analysis, explanation = self._analyze_map_prompt_categorization(map_prompt)
+
+            if analysis:
+                dedup = False
+        else:
+            self.console.log(
+                "[yellow]No map prompt found in intermediates for analysis.[/yellow]"
+            )
+
+        # TODO: figure out why this would ever be the case
+        if not map_prompt:
+            map_prompt = "N/A"
+
+        if dedup is False:
+            dedup, explanation = self._determine_duplicate_keys(
+                input_data, reduce_key, map_prompt
+            )
+
+        # Now do the last attempt of pairwise comparisons
+        if dedup is False:
+            # Sample up to 20 random pairs of keys for duplicate analysis
+            sampled_pairs = self._sample_random_pairs(input_data, 20)
+
+            # Use LLM to check for duplicates
+            duplicates_found, explanation = self._check_duplicates_with_llm(
+                input_data, sampled_pairs, reduce_key, map_prompt
+            )
+
+            if duplicates_found:
+                dedup = True
+
+        return dedup, explanation
+
+    return False, ""
+
+
+
+ +
+ + + +
+ +
+ +
+ + + + + + + + + + + + + +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/api-reference/python/index.html b/api-reference/python/index.html new file mode 100644 index 00000000..7d761d2e --- /dev/null +++ b/api-reference/python/index.html @@ -0,0 +1,3260 @@ + + + + + + + + + + + + + + + + + + + + + + + Python API - docetl docs + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + +
+ + + + + + + +
+ +
+ + + + +
+
+ + + +
+
+
+ + + + + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + + + + +

Python API

+

Operations

+ + +
+ + + +

+ docetl.schemas.MapOp = map.MapOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.ResolveOp = resolve.ResolveOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.ReduceOp = reduce.ReduceOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.ParallelMapOp = map.ParallelMapOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.FilterOp = filter.FilterOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.EquijoinOp = equijoin.EquijoinOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.SplitOp = split.SplitOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.GatherOp = gather.GatherOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.UnnestOp = unnest.UnnestOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.SampleOp = sample.SampleOperation.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.ClusterOp = cluster.ClusterOperation.schema + + + module-attribute + + +

+ + +
+
+ +

Dataset and Pipeline

+ + +
+ + + +

+ docetl.schemas.Dataset = dataset.Dataset.schema + + + module-attribute + + +

+ + +
+
+ +
+ +
+ + + +

+ docetl.schemas.ParsingTool + + +

+ + +
+

+ Bases: BaseModel

+ + +

Represents a parsing tool used for custom data parsing in the pipeline.

+ + +

Attributes:

+ + + + + + + + + + + + + + + + + + + + +
NameTypeDescription
name + str + +
+

The name of the parsing tool. This should be unique within the pipeline configuration.

+
+
function_code + str + +
+

The Python code defining the parsing function. This code will be executed + to parse the input data according to the specified logic. It should return a list of strings, where each string is its own document.

+
+
+ + +
+ Example +
parsing_tools:
+  - name: ocr_parser
+    function_code: |
+      import pytesseract
+      from pdf2image import convert_from_path
+      def ocr_parser(filename: str) -> List[str]:
+          images = convert_from_path(filename)
+          text = ""
+          for image in images:
+              text += pytesseract.image_to_string(image)
+          return [text]
+
+
+ + + + + +
+ Source code in docetl/base_schemas.py +
19
+20
+21
+22
+23
+24
+25
+26
+27
+28
+29
+30
+31
+32
+33
+34
+35
+36
+37
+38
+39
+40
+41
+42
+43
+44
+45
class ParsingTool(BaseModel):
+    """
+    Represents a parsing tool used for custom data parsing in the pipeline.
+
+    Attributes:
+        name (str): The name of the parsing tool. This should be unique within the pipeline configuration.
+        function_code (str): The Python code defining the parsing function. This code will be executed
+                             to parse the input data according to the specified logic. It should return a list of strings, where each string is its own document.
+
+    Example:
+        ```yaml
+        parsing_tools:
+          - name: ocr_parser
+            function_code: |
+              import pytesseract
+              from pdf2image import convert_from_path
+              def ocr_parser(filename: str) -> List[str]:
+                  images = convert_from_path(filename)
+                  text = ""
+                  for image in images:
+                      text += pytesseract.image_to_string(image)
+                  return [text]
+        ```
+    """
+
+    name: str
+    function_code: str
+
+
+ + + +
+ + + + + + + + + + + +
+ +
+ +
+ +
+ + + +

+ docetl.schemas.PipelineStep + + +

+ + +
+

+ Bases: BaseModel

+ + +

Represents a step in the pipeline.

+ + +

Attributes:

+ + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescription
name + str + +
+

The name of the step.

+
+
operations + List[Union[Dict[str, Any], str]] + +
+

A list of operations to be applied in this step. +Each operation can be either a string (the name of the operation) or a dictionary +(for more complex configurations).

+
+
input + Optional[str] + +
+

The input for this step. It can be either the name of a dataset +or the name of a previous step. If not provided, the step will use the output +of the previous step as its input.

+
+
+ + +
+ Example +
# Simple step with a single operation
+process_step = PipelineStep(
+    name="process_step",
+    input="my_dataset",
+    operations=["process"]
+)
+
+# Step with multiple operations
+summarize_step = PipelineStep(
+    name="summarize_step",
+    input="process_step",
+    operations=["summarize"]
+)
+
+# Step with a more complex operation configuration
+custom_step = PipelineStep(
+    name="custom_step",
+    input="previous_step",
+    operations=[
+        {
+            "custom_operation": {
+                "model": "gpt-4",
+                "prompt": "Perform a custom analysis on the following text:"
+            }
+        }
+    ]
+)
+
+

These examples show different ways to configure pipeline steps, from simple +single-operation steps to more complex configurations with custom parameters.

+ + + + + + +
+ Source code in docetl/base_schemas.py +
48
+49
+50
+51
+52
+53
+54
+55
+56
+57
+58
+59
+60
+61
+62
+63
+64
+65
+66
+67
+68
+69
+70
+71
+72
+73
+74
+75
+76
+77
+78
+79
+80
+81
+82
+83
+84
+85
+86
+87
+88
+89
+90
+91
+92
+93
+94
+95
+96
+97
+98
class PipelineStep(BaseModel):
+    """
+    Represents a step in the pipeline.
+
+    Attributes:
+        name (str): The name of the step.
+        operations (List[Union[Dict[str, Any], str]]): A list of operations to be applied in this step.
+            Each operation can be either a string (the name of the operation) or a dictionary
+            (for more complex configurations).
+        input (Optional[str]): The input for this step. It can be either the name of a dataset
+            or the name of a previous step. If not provided, the step will use the output
+            of the previous step as its input.
+
+    Example:
+        ```python
+        # Simple step with a single operation
+        process_step = PipelineStep(
+            name="process_step",
+            input="my_dataset",
+            operations=["process"]
+        )
+
+        # Step with multiple operations
+        summarize_step = PipelineStep(
+            name="summarize_step",
+            input="process_step",
+            operations=["summarize"]
+        )
+
+        # Step with a more complex operation configuration
+        custom_step = PipelineStep(
+            name="custom_step",
+            input="previous_step",
+            operations=[
+                {
+                    "custom_operation": {
+                        "model": "gpt-4",
+                        "prompt": "Perform a custom analysis on the following text:"
+                    }
+                }
+            ]
+        )
+        ```
+
+    These examples show different ways to configure pipeline steps, from simple
+    single-operation steps to more complex configurations with custom parameters.
+    """
+
+    name: str
+    operations: List[Union[Dict[str, Any], str]]
+    input: Optional[str] = None
+
+
+ + + +
+ + + + + + + + + + + +
+ +
+ +
+ +
+ + + +

+ docetl.schemas.PipelineOutput + + +

+ + +
+

+ Bases: BaseModel

+ + +

Represents the output configuration for a pipeline.

+ + +

Attributes:

+ + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescription
type + str + +
+

The type of output. This could be 'file', 'database', etc.

+
+
path + str + +
+

The path where the output will be stored. This could be a file path, + database connection string, etc., depending on the type.

+
+
intermediate_dir + Optional[str] + +
+

The directory to store intermediate results, + if applicable. Defaults to None.

+
+
+ + +
+ Example +
output = PipelineOutput(
+    type="file",
+    path="/path/to/output.json",
+    intermediate_dir="/path/to/intermediate/results"
+)
+
+
+ + + + + +
+ Source code in docetl/base_schemas.py +
100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
class PipelineOutput(BaseModel):
+    """
+    Represents the output configuration for a pipeline.
+
+    Attributes:
+        type (str): The type of output. This could be 'file', 'database', etc.
+        path (str): The path where the output will be stored. This could be a file path,
+                    database connection string, etc., depending on the type.
+        intermediate_dir (Optional[str]): The directory to store intermediate results,
+                                          if applicable. Defaults to None.
+
+    Example:
+        ```python
+        output = PipelineOutput(
+            type="file",
+            path="/path/to/output.json",
+            intermediate_dir="/path/to/intermediate/results"
+        )
+        ```
+    """
+
+    type: str
+    path: str
+    intermediate_dir: Optional[str] = None
+
+
+ + + +
+ + + + + + + + + + + +
+ +
+ +
+ +
+ + + +

+ docetl.api.Pipeline + + +

+ + +
+ + +

Represents a complete document processing pipeline.

+ + +

Attributes:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescription
name + str + +
+

The name of the pipeline.

+
+
datasets + Dict[str, Dataset] + +
+

A dictionary of datasets used in the pipeline, + where keys are dataset names and values are Dataset objects.

+
+
operations + List[OpType] + +
+

A list of operations to be performed in the pipeline.

+
+
steps + List[PipelineStep] + +
+

A list of steps that make up the pipeline.

+
+
output + PipelineOutput + +
+

The output configuration for the pipeline.

+
+
parsing_tools + List[ParsingTool] + +
+

A list of parsing tools used in the pipeline. + Defaults to an empty list.

+
+
default_model + Optional[str] + +
+

The default language model to use for operations + that require one. Defaults to None.

+
+
+ + +
+ Example +
def custom_parser(text: str) -> List[str]:
+    # this will convert the text in the column to uppercase
+    # You should return a list of strings, where each string is a separate document
+    return [text.upper()]
+
+pipeline = Pipeline(
+    name="document_processing_pipeline",
+    datasets={
+        "input_data": Dataset(type="file", path="/path/to/input.json", parsing=[{"name": "custom_parser", "input_key": "content", "output_key": "uppercase_content"}]),
+    },
+    parsing_tools=[custom_parser],
+    operations=[
+        MapOp(
+            name="process",
+            type="map",
+            prompt="Determine what type of document this is: {{ input.uppercase_content }}",
+            output={"schema": {"document_type": "string"}}
+        ),
+        ReduceOp(
+            name="summarize",
+            type="reduce",
+            reduce_key="document_type",
+            prompt="Summarize the processed contents: {% for item in inputs %}{{ item.uppercase_content }} {% endfor %}",
+            output={"schema": {"summary": "string"}}
+        )
+    ],
+    steps=[
+        PipelineStep(name="process_step", input="input_data", operations=["process"]),
+        PipelineStep(name="summarize_step", input="process_step", operations=["summarize"])
+    ],
+    output=PipelineOutput(type="file", path="/path/to/output.json"),
+    default_model="gpt-4o-mini"
+)
+
+

This example shows a complete pipeline configuration with datasets, operations, +steps, and output settings.

+ + + + + + +
+ Source code in docetl/api.py +
 81
+ 82
+ 83
+ 84
+ 85
+ 86
+ 87
+ 88
+ 89
+ 90
+ 91
+ 92
+ 93
+ 94
+ 95
+ 96
+ 97
+ 98
+ 99
+100
+101
+102
+103
+104
+105
+106
+107
+108
+109
+110
+111
+112
+113
+114
+115
+116
+117
+118
+119
+120
+121
+122
+123
+124
+125
+126
+127
+128
+129
+130
+131
+132
+133
+134
+135
+136
+137
+138
+139
+140
+141
+142
+143
+144
+145
+146
+147
+148
+149
+150
+151
+152
+153
+154
+155
+156
+157
+158
+159
+160
+161
+162
+163
+164
+165
+166
+167
+168
+169
+170
+171
+172
+173
+174
+175
+176
+177
+178
+179
+180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
+219
+220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
+239
+240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
+255
+256
+257
+258
+259
+260
+261
+262
+263
+264
+265
+266
+267
+268
+269
+270
+271
+272
+273
+274
+275
+276
+277
+278
+279
+280
+281
+282
+283
+284
+285
+286
+287
+288
+289
+290
+291
+292
+293
+294
+295
+296
+297
+298
+299
+300
+301
+302
+303
+304
+305
+306
+307
+308
+309
+310
+311
+312
+313
+314
+315
+316
+317
+318
+319
+320
+321
+322
+323
+324
+325
+326
+327
+328
+329
+330
+331
+332
+333
+334
+335
+336
+337
class Pipeline:
+    """
+    Represents a complete document processing pipeline.
+
+    Attributes:
+        name (str): The name of the pipeline.
+        datasets (Dict[str, Dataset]): A dictionary of datasets used in the pipeline,
+                                       where keys are dataset names and values are Dataset objects.
+        operations (List[OpType]): A list of operations to be performed in the pipeline.
+        steps (List[PipelineStep]): A list of steps that make up the pipeline.
+        output (PipelineOutput): The output configuration for the pipeline.
+        parsing_tools (List[ParsingTool]): A list of parsing tools used in the pipeline.
+                                           Defaults to an empty list.
+        default_model (Optional[str]): The default language model to use for operations
+                                       that require one. Defaults to None.
+
+    Example:
+        ```python
+        def custom_parser(text: str) -> List[str]:
+            # this will convert the text in the column to uppercase
+            # You should return a list of strings, where each string is a separate document
+            return [text.upper()]
+
+        pipeline = Pipeline(
+            name="document_processing_pipeline",
+            datasets={
+                "input_data": Dataset(type="file", path="/path/to/input.json", parsing=[{"name": "custom_parser", "input_key": "content", "output_key": "uppercase_content"}]),
+            },
+            parsing_tools=[custom_parser],
+            operations=[
+                MapOp(
+                    name="process",
+                    type="map",
+                    prompt="Determine what type of document this is: {{ input.uppercase_content }}",
+                    output={"schema": {"document_type": "string"}}
+                ),
+                ReduceOp(
+                    name="summarize",
+                    type="reduce",
+                    reduce_key="document_type",
+                    prompt="Summarize the processed contents: {% for item in inputs %}{{ item.uppercase_content }} {% endfor %}",
+                    output={"schema": {"summary": "string"}}
+                )
+            ],
+            steps=[
+                PipelineStep(name="process_step", input="input_data", operations=["process"]),
+                PipelineStep(name="summarize_step", input="process_step", operations=["summarize"])
+            ],
+            output=PipelineOutput(type="file", path="/path/to/output.json"),
+            default_model="gpt-4o-mini"
+        )
+        ```
+
+    This example shows a complete pipeline configuration with datasets, operations,
+    steps, and output settings.
+    """
+
+    def __init__(
+        self,
+        name: str,
+        datasets: Dict[str, Dataset],
+        operations: List[OpType],
+        steps: List[PipelineStep],
+        output: PipelineOutput,
+        parsing_tools: List[Union[ParsingTool, Callable]] = [],
+        default_model: Optional[str] = None,
+        rate_limits: Optional[Dict[str, int]] = None,
+    ):
+        self.name = name
+        self.datasets = datasets
+        self.operations = operations
+        self.steps = steps
+        self.output = output
+        self.parsing_tools = [
+            (
+                tool
+                if isinstance(tool, ParsingTool)
+                else ParsingTool(
+                    name=tool.__name__, function_code=inspect.getsource(tool)
+                )
+            )
+            for tool in parsing_tools
+        ]
+        self.default_model = default_model
+        self.rate_limits = rate_limits
+        self._load_env()
+
+    def _load_env(self):
+        from dotenv import load_dotenv
+        import os
+
+        # Get the current working directory
+        cwd = os.getcwd()
+
+        # Load .env file from the current working directory if it exists
+        env_file = os.path.join(cwd, ".env")
+        if os.path.exists(env_file):
+            load_dotenv(env_file)
+
+    def optimize(
+        self,
+        max_threads: Optional[int] = None,
+        model: str = "gpt-4o",
+        resume: bool = False,
+        timeout: int = 60,
+    ) -> "Pipeline":
+        """
+        Optimize the pipeline using the Optimizer.
+
+        Args:
+            max_threads (Optional[int]): Maximum number of threads to use for optimization.
+            model (str): The model to use for optimization. Defaults to "gpt-4o".
+            resume (bool): Whether to resume optimization from a previous state. Defaults to False.
+            timeout (int): Timeout for optimization in seconds. Defaults to 60.
+
+        Returns:
+            Pipeline: An optimized version of the pipeline.
+        """
+        config = self._to_dict()
+        runner = DSLRunner(
+            config,
+            base_name=os.path.join(os.getcwd(), self.name),
+            yaml_file_suffix=self.name,
+            max_threads=max_threads,
+        )
+        optimized_config, _ = runner.optimize(return_pipeline=False)
+
+        updated_pipeline = Pipeline(
+            name=self.name,
+            datasets=self.datasets,
+            operations=self.operations,
+            steps=self.steps,
+            output=self.output,
+            default_model=self.default_model,
+            parsing_tools=self.parsing_tools,
+        )
+        updated_pipeline._update_from_dict(optimized_config)
+        return updated_pipeline
+
+    def run(self, max_threads: Optional[int] = None) -> float:
+        """
+        Run the pipeline using the DSLRunner.
+
+        Args:
+            max_threads (Optional[int]): Maximum number of threads to use for execution.
+
+        Returns:
+            float: The total cost of running the pipeline.
+        """
+        config = self._to_dict()
+        runner = DSLRunner(
+            config,
+            base_name=os.path.join(os.getcwd(), self.name),
+            yaml_file_suffix=self.name,
+            max_threads=max_threads,
+        )
+        result = runner.load_run_save()
+        return result
+
+    def to_yaml(self, path: str) -> None:
+        """
+        Convert the Pipeline object to a YAML string and save it to a file.
+
+        Args:
+            path (str): Path to save the YAML file.
+
+        Returns:
+            None
+        """
+        config = self._to_dict()
+        with open(path, "w") as f:
+            yaml.safe_dump(config, f)
+
+        print(f"[green]Pipeline saved to {path}[/green]")
+
+    def _to_dict(self) -> Dict[str, Any]:
+        """
+        Convert the Pipeline object to a dictionary representation.
+
+        Returns:
+            Dict[str, Any]: Dictionary representation of the Pipeline.
+        """
+        d = {
+            "datasets": {
+                name: dataset.dict() for name, dataset in self.datasets.items()
+            },
+            "operations": [
+                {k: v for k, v in op.dict().items() if v is not None}
+                for op in self.operations
+            ],
+            "pipeline": {
+                "steps": [
+                    {k: v for k, v in step.dict().items() if v is not None}
+                    for step in self.steps
+                ],
+                "output": self.output.dict(),
+            },
+            "default_model": self.default_model,
+            "parsing_tools": (
+                [tool.dict() for tool in self.parsing_tools]
+                if self.parsing_tools
+                else None
+            ),
+        }
+        if self.rate_limits:
+            d["rate_limits"] = self.rate_limits
+        return d
+
+    def _update_from_dict(self, config: Dict[str, Any]):
+        """
+        Update the Pipeline object from a dictionary representation.
+
+        Args:
+            config (Dict[str, Any]): Dictionary representation of the Pipeline.
+        """
+        self.datasets = {
+            name: Dataset(
+                type=dataset["type"],
+                source=dataset["source"],
+                path=dataset["path"],
+                parsing=dataset.get("parsing"),
+            )
+            for name, dataset in config["datasets"].items()
+        }
+        self.operations = []
+        for op in config["operations"]:
+            op_type = op.pop("type")
+            if op_type == "map":
+                self.operations.append(MapOp(**op, type=op_type))
+            elif op_type == "resolve":
+                self.operations.append(ResolveOp(**op, type=op_type))
+            elif op_type == "reduce":
+                self.operations.append(ReduceOp(**op, type=op_type))
+            elif op_type == "parallel_map":
+                self.operations.append(ParallelMapOp(**op, type=op_type))
+            elif op_type == "filter":
+                self.operations.append(FilterOp(**op, type=op_type))
+            elif op_type == "equijoin":
+                self.operations.append(EquijoinOp(**op, type=op_type))
+            elif op_type == "split":
+                self.operations.append(SplitOp(**op, type=op_type))
+            elif op_type == "gather":
+                self.operations.append(GatherOp(**op, type=op_type))
+            elif op_type == "unnest":
+                self.operations.append(UnnestOp(**op, type=op_type))
+            elif op_type == "cluster":
+                self.operations.append(ClusterOp(**op, type=op_type))
+            elif op_type == "sample":
+                self.operations.append(SampleOp(**op, type=op_type))
+        self.steps = [PipelineStep(**step) for step in config["pipeline"]["steps"]]
+        self.output = PipelineOutput(**config["pipeline"]["output"])
+        self.default_model = config.get("default_model")
+        self.parsing_tools = (
+            [ParsingTool(**tool) for tool in config.get("parsing_tools", [])]
+            if config.get("parsing_tools")
+            else []
+        )
+
+
+ + + +
+ + + + + + + + + +
+ + +

+ optimize(max_threads=None, model='gpt-4o', resume=False, timeout=60) + +

+ + +
+ +

Optimize the pipeline using the Optimizer.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ max_threads + + Optional[int] + +
+

Maximum number of threads to use for optimization.

+
+
+ None +
+ model + + str + +
+

The model to use for optimization. Defaults to "gpt-4o".

+
+
+ 'gpt-4o' +
+ resume + + bool + +
+

Whether to resume optimization from a previous state. Defaults to False.

+
+
+ False +
+ timeout + + int + +
+

Timeout for optimization in seconds. Defaults to 60.

+
+
+ 60 +
+ + +

Returns:

+ + + + + + + + + + + + + +
Name TypeDescription
Pipeline + Pipeline + +
+

An optimized version of the pipeline.

+
+
+ +
+ Source code in docetl/api.py +
180
+181
+182
+183
+184
+185
+186
+187
+188
+189
+190
+191
+192
+193
+194
+195
+196
+197
+198
+199
+200
+201
+202
+203
+204
+205
+206
+207
+208
+209
+210
+211
+212
+213
+214
+215
+216
+217
+218
def optimize(
+    self,
+    max_threads: Optional[int] = None,
+    model: str = "gpt-4o",
+    resume: bool = False,
+    timeout: int = 60,
+) -> "Pipeline":
+    """
+    Optimize the pipeline using the Optimizer.
+
+    Args:
+        max_threads (Optional[int]): Maximum number of threads to use for optimization.
+        model (str): The model to use for optimization. Defaults to "gpt-4o".
+        resume (bool): Whether to resume optimization from a previous state. Defaults to False.
+        timeout (int): Timeout for optimization in seconds. Defaults to 60.
+
+    Returns:
+        Pipeline: An optimized version of the pipeline.
+    """
+    config = self._to_dict()
+    runner = DSLRunner(
+        config,
+        base_name=os.path.join(os.getcwd(), self.name),
+        yaml_file_suffix=self.name,
+        max_threads=max_threads,
+    )
+    optimized_config, _ = runner.optimize(return_pipeline=False)
+
+    updated_pipeline = Pipeline(
+        name=self.name,
+        datasets=self.datasets,
+        operations=self.operations,
+        steps=self.steps,
+        output=self.output,
+        default_model=self.default_model,
+        parsing_tools=self.parsing_tools,
+    )
+    updated_pipeline._update_from_dict(optimized_config)
+    return updated_pipeline
+
+
+
+ +
+ +
+ + +

+ run(max_threads=None) + +

+ + +
+ +

Run the pipeline using the DSLRunner.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ max_threads + + Optional[int] + +
+

Maximum number of threads to use for execution.

+
+
+ None +
+ + +

Returns:

+ + + + + + + + + + + + + +
Name TypeDescription
float + float + +
+

The total cost of running the pipeline.

+
+
+ +
+ Source code in docetl/api.py +
220
+221
+222
+223
+224
+225
+226
+227
+228
+229
+230
+231
+232
+233
+234
+235
+236
+237
+238
def run(self, max_threads: Optional[int] = None) -> float:
+    """
+    Run the pipeline using the DSLRunner.
+
+    Args:
+        max_threads (Optional[int]): Maximum number of threads to use for execution.
+
+    Returns:
+        float: The total cost of running the pipeline.
+    """
+    config = self._to_dict()
+    runner = DSLRunner(
+        config,
+        base_name=os.path.join(os.getcwd(), self.name),
+        yaml_file_suffix=self.name,
+        max_threads=max_threads,
+    )
+    result = runner.load_run_save()
+    return result
+
+
+
+ +
+ +
+ + +

+ to_yaml(path) + +

+ + +
+ +

Convert the Pipeline object to a YAML string and save it to a file.

+ + +

Parameters:

+ + + + + + + + + + + + + + + + + +
NameTypeDescriptionDefault
+ path + + str + +
+

Path to save the YAML file.

+
+
+ required +
+ + +

Returns:

+ + + + + + + + + + + + + +
TypeDescription
+ None + +
+

None

+
+
+ +
+ Source code in docetl/api.py +
240
+241
+242
+243
+244
+245
+246
+247
+248
+249
+250
+251
+252
+253
+254
def to_yaml(self, path: str) -> None:
+    """
+    Convert the Pipeline object to a YAML string and save it to a file.
+
+    Args:
+        path (str): Path to save the YAML file.
+
+    Returns:
+        None
+    """
+    config = self._to_dict()
+    with open(path, "w") as f:
+        yaml.safe_dump(config, f)
+
+    print(f"[green]Pipeline saved to {path}[/green]")
+
+
+
+ +
+ + + +
+ +
+ +
+ + + + + + + + + + + + + +
+
+ + + +
+ + + +
+ + + +
+
+
+
+ + + + + + + + + + \ No newline at end of file diff --git a/assets/_mkdocstrings.css b/assets/_mkdocstrings.css new file mode 100644 index 00000000..b500381b --- /dev/null +++ b/assets/_mkdocstrings.css @@ -0,0 +1,143 @@ + +/* Avoid breaking parameter names, etc. in table cells. */ +.doc-contents td code { + word-break: normal !important; +} + +/* No line break before first paragraph of descriptions. */ +.doc-md-description, +.doc-md-description>p:first-child { + display: inline; +} + +/* Max width for docstring sections tables. */ +.doc .md-typeset__table, +.doc .md-typeset__table table { + display: table !important; + width: 100%; +} + +.doc .md-typeset__table tr { + display: table-row; +} + +/* Defaults in Spacy table style. */ +.doc-param-default { + float: right; +} + +/* Parameter headings must be inline, not blocks. */ +.doc-heading-parameter { + display: inline; +} + +/* Prefer space on the right, not the left of parameter permalinks. */ +.doc-heading-parameter .headerlink { + margin-left: 0 !important; + margin-right: 0.2rem; +} + +/* Backward-compatibility: docstring section titles in bold. */ +.doc-section-title { + font-weight: bold; +} + +/* Symbols in Navigation and ToC. */ +:root, :host, +[data-md-color-scheme="default"] { + --doc-symbol-parameter-fg-color: #df50af; + --doc-symbol-attribute-fg-color: #953800; + --doc-symbol-function-fg-color: #8250df; + --doc-symbol-method-fg-color: #8250df; + --doc-symbol-class-fg-color: #0550ae; + --doc-symbol-module-fg-color: #5cad0f; + + --doc-symbol-parameter-bg-color: #df50af1a; + --doc-symbol-attribute-bg-color: #9538001a; + --doc-symbol-function-bg-color: #8250df1a; + --doc-symbol-method-bg-color: #8250df1a; + --doc-symbol-class-bg-color: #0550ae1a; + --doc-symbol-module-bg-color: #5cad0f1a; +} + +[data-md-color-scheme="slate"] { + --doc-symbol-parameter-fg-color: #ffa8cc; + --doc-symbol-attribute-fg-color: #ffa657; + --doc-symbol-function-fg-color: #d2a8ff; + --doc-symbol-method-fg-color: #d2a8ff; + --doc-symbol-class-fg-color: #79c0ff; + --doc-symbol-module-fg-color: #baff79; + + --doc-symbol-parameter-bg-color: #ffa8cc1a; + --doc-symbol-attribute-bg-color: #ffa6571a; + --doc-symbol-function-bg-color: #d2a8ff1a; + --doc-symbol-method-bg-color: #d2a8ff1a; + --doc-symbol-class-bg-color: #79c0ff1a; + --doc-symbol-module-bg-color: #baff791a; +} + +code.doc-symbol { + border-radius: .1rem; + font-size: .85em; + padding: 0 .3em; + font-weight: bold; +} + +code.doc-symbol-parameter { + color: var(--doc-symbol-parameter-fg-color); + background-color: var(--doc-symbol-parameter-bg-color); +} + +code.doc-symbol-parameter::after { + content: "param"; +} + +code.doc-symbol-attribute { + color: var(--doc-symbol-attribute-fg-color); + background-color: var(--doc-symbol-attribute-bg-color); +} + +code.doc-symbol-attribute::after { + content: "attr"; +} + +code.doc-symbol-function { + color: var(--doc-symbol-function-fg-color); + background-color: var(--doc-symbol-function-bg-color); +} + +code.doc-symbol-function::after { + content: "func"; +} + +code.doc-symbol-method { + color: var(--doc-symbol-method-fg-color); + background-color: var(--doc-symbol-method-bg-color); +} + +code.doc-symbol-method::after { + content: "meth"; +} + +code.doc-symbol-class { + color: var(--doc-symbol-class-fg-color); + background-color: var(--doc-symbol-class-bg-color); +} + +code.doc-symbol-class::after { + content: "class"; +} + +code.doc-symbol-module { + color: var(--doc-symbol-module-fg-color); + background-color: var(--doc-symbol-module-bg-color); +} + +code.doc-symbol-module::after { + content: "mod"; +} + +.doc-signature .autorefs { + color: inherit; + border-bottom: 1px dotted currentcolor; +} diff --git a/assets/docetl-favicon-color.png b/assets/docetl-favicon-color.png new file mode 100644 index 0000000000000000000000000000000000000000..51fe742f01c4f74fa5fb80765e979e64f964d90f GIT binary patch literal 1188 zcmV;V1Y7%wP)C0000RP)t-sOlfgC zexW*kpgMh^I(?x!eV{shp*nq`JAR>Fi8EvX000PdQchCt)Yjho&C@000Sa zNLh0L00Za%00Za&wR=))000CVNklKJ>06>Ia6pN@pW|CIWZ%9@kTZ{b`160_qcn`iyHtkj| zMg#z=xIj}%c7t$O%N8S0JX|TIlwh!MT8loBJcIKejZ*?rrEw!|m>I@x0anrpnh$s$ z#6zW9yYNw4z_3V{rfpwCQj(Z>dP{dX*1QE67cYGqU(XQJflWpiai|v2fo7}g*ZXb` z_Y4xXpm`Fn4q!+;qm;-ul9T9wE%46)3GyaHw}9P7Aep$t1FAU}p~ZF}?#sbxg;peU z+?S&-E#Y~T#dYAbLY-GWKnLcUf@aIX_&ESqJW4a`0Y;#nHDF5fr}#lck1y2Oa9HI{ z#SPCffIMc&z@v9Oa~uOoG6^?|TOm0%k>(r}^9Dw(&~ucnWMXCM<_)KBs6&gCQ?ZFQ z^=fdpT8}TG^v;5zON03iAsd}}d>N78Zk5Cb`9nXH0QC~@7-D1gjNk= zvD~8zjKl~?+Xj{#_W1r3JV@3;fbBr{3n&&(sDXbD8GGFiBVY~2KY6WVKrFk5D*`BQ zuXPIny%?Ua9vGUTIXO85T%Ynd0Y3dD(?{(OA;2F9uq3;7)EFj!XEV1ssS71u;*(7zktSlNOV8;ZsRzH+d)@=l!LH-A0h!UWq$+4EdQpOzwX!jjU zC4WWc1_E$Ez-ogt2)Hs|G=f_G$47$^1awc}BKiy`AiJ+#-z0$13-W(YfTMMyUnii^ z2~w~|;kh0+g~NO6WW86DKTUvwMG}e=uqc735y|fo&~Qo>B!EFjV$o6`28}52bT@&6 zTbN5Hk@2s1l_^?sx6lNom81GmN$3mZ<{w}#U(KIcr&`*p4-`5Eg+Li!qzfFS**Cnp z)T8q;($Wc%s|ngo07F(-2+Yq9y1WYafkS2t+cpsDcPrvIg5}|-YQ!+t3ySv3W?lNe z){PhreL}AhFitji)B=OPcSLEI4fL`12psCdH5<|__yxM{e0!uIu2zH}D-C721 z1akT{ccKpHvX-hGZPL8&r4@BZcQ01i?Ug&N`k(#%n8%O$bLQ2PVPqk2jnE0000G%OaQA*nj6=4B43 z$sMF*GQ0Avojt&u8VBdSeLJXhjoBNmnnbvsb92!KnpL|tnGjdgd?Y2`>*J6k+9?H4 z$nG^i2bvGTTW3j6UlN4az9Cln!4#XNd{hIFm@w>p)fUzrxZ~q+?jWIiNGCg$higAB z?X!6FIkk%mIrN-%>dIE_z_zh>Cnsy2NOG#UnNY1RfGACH%60m&g_n05YyPl}+QBYs zw9=#xIe^(RUpQ4Egu**JwM~_lIj4Z*#M@*K%q0=*JZLdF=Aib=o;^BHeR9$~yMV*; z;p}*9>iFPxd+@`pNcFR*9QRc7-gHJaS^fca;;CWmjKc{<+2N=z`BW(l>sz!5VM`Pq zOT3q1boCgENVaE#jX!*V`KUU7B)EC`86 zVGd<##o7nWqJ`7o2*Lx9w3x3noy=qRStNu4k$h=)W0j131TV%3&qk1g9gaUmE11XACBAc5Xtb6BMWKrh!37fatHC0%FC3H)(kugw=;mXxx_z^eZ+opyu*H$J zVtm$60(eZZ7W!1y$gfdeq8>$HtW%t@Vi%U`#B=hdbHyiF7o+G3RtBVFw$STGzL%RT zd8_lYl5du7mhph?fQ9;Xd9jnDUNg%Q4PDYs;n!*KG}|8c1(Pq%4eMCYmfc;e1z z!3tmkrj$R)Rck73p9!3qo8g`5nyItJ6bcf;&B+jw5F!)euyr5lOlKQD7!$HvYy{g? z+i8#e$a+_*!$qDh_k_DF(mUHbQHVVz?Q7O~%H}B9@XxdkPB9H7UB5~V?wC(;p9+F< zB$<;l#8MnaI7TSO7&4AC3v^QS>J4Ibh7BqW#H$4=W1r3%+H#e1W^&UTc2-`tI=j?0aN*`?TgaTKsredH^s zAyy;S=@KbgOv6U=Q}B};F3mC}2el#1s!+J5tL8O(utpHM4f!%Z`NLX^m#gWvqXH8` zZX$F-dLrk-V`46%!7YmZYU8Ot)jwsv6#XcuEx76d8-6_5JH)(N{Cy*_tbuK)ap|m< zY;p7}_tg7|gO-*!(W#(1KM>yB;HrI-X2k@BiImA3x3r&a?`q9z%o4wce3F)lHGd3p zPHymTnrZ^rz^8lWnZB;K^%W$QCygY<&<$l$Kiua14)Xn`wjB5K*{;Z*hEI|oY&!6Z z(pc2br^jcr`a>Dupl-i(tA15$TB}ZPVSkeLciy?yk^&2YQY*Sg=0{3L-|4;S*Xi}? zwWV8RNTg|GI67TB=Y#wLY&t`_J_fHkR5gj$pwCApznhXQQ2)%jsrGq}~XJ!>EXh#C_qpcPsF* zex_b!J$ilPWD+wccAE%~Nbk+j8@o3Vy6VOAO5bA*;&xO%4y;)-TjTd@@SPjA`YaB6 z=Wa)h9e)rodegPB^duvLi)X4f)p)?%kC~i=hALS6m#|cd*;Ial&A?p0re}7ZSssO{ zp}T~C-fln7X5ygKK-NGqn>gE{Dy6osT8-+<5r?zt>$xWWWk8jjFY<)wCLsY^J?!Hvg+^A zu+$hlw!rK@eQABZ5x+6TmX`_!7(sO)eIGxZnslhUsb9JGxs$kdeEe>ATO{+sC6Lw+?_Io57`ihrZZC<(z9(&Qjm^ zuN@119(o^w+`N8PFhBsm76Rv5X2A=FhkZw5OINPIepqe69zc{#XQ4v}LDbpoPv3Q!H*AHf+|iufLra*Z@EKuUrM}V%@}* zrv-7iXz6N`waD?hJZFDSVuG<;yd0!;n|I{u?+Unly~vL-<@l~jw%W7xg7(z3r`==M zXHxqGrj)t4YMa1H`}_0hfy&qZ{0}S+J|6}{4#=i>C^{(3cIza~WPB#pm-fLI;16Q1 z9?dJw?STh|xfa@f%P&5zGEi36IBHvM_s;UWQnrw@nCdzSzPrj= zWjNgZl`NwgcyePk-?y(qgD!xolx)XfD6Jd(6rMZnF%i)k;SjSD<1ULAa^sjS4Bfa& zSOp}LO0xi}Dd%Y2D4S@dgpVXbZy}AlEc}uD__s^0v#)y>2UBfR85Rkjn}L^ik|(^E z%VJC>{rMRvoVGouvF-srgh<{ZNVCgGgjo|#`^FJF7x0JS@P415;CWYM$HyxP7xJv4 zj^Dx@2@Ms|{TK!NW7<)^jXr7I@jI?jE)WFhDb_h=FE8g1JirSBF6D*7$5!AMa!)Tk zU>2hfkbbf%#%CjTt_F5SYWA9%NZg2POeEwOCnPk)6*A&OjrbrTp{7P7VIckzBR-0S zD1XJGD-@#sb&b+>pHN;;K}`+uSI@@V&d%M((Ze^RRG=AA)PmD9BVQv;4M`ghH$E#{ z4{JNV05`;L8Sx_>Ac?qiv-7oL3~+OG_mK>cVg8*$5^;SW%+JjDJBhE046~7@4x@sH zw;kgnJ^?-fW?5`TMn-9GTYJf;ipu}Wj`&Z8+0ob6Q<9(G-`}6lUzpFs+kszDLPCOH zK!{&Rh!>H9*C){3*D8S5-G}8*A%B&lXy;?&?d0j}>(?!!O7u!2j>E5m}}0LnU>b0_-2?+l^>))dO$E^B3cHRmeZiqU4W&d+){+0Rfk^jml&41tZzYWEo zasC~O7-(5+Y5srTG+FH6CR!>aBsnBCMfqm|$cJ56ZFWk2k4&##h)rmq;-PR`3m_*J zD+}tzZ5D9u2ol9Op0N^HYkVGBdLzzT>*N+k?FnQJ$D#a3=+BJ30Tw5j{t4(H5jTNi;Bh}uQFPsj&EG} zEx{3>o8?aoi^JD+L8j+p_gPFn!n2hc^hdXBl$+dSS#em1as$gRJ7g}4rqmX z`Gy6#mr`b5%gzH7I8u;H696^|M?szfFlzF6tSf*F{WxH_Nz^*UCWFU5{$a_M7srmp zeZsPA;5&wK5E(`&%>GWvTkhb{Z=PW`hgzsou(%P%$*yZ>1nj{(0whrM;Fgje0Yogf zkQi$~SCKbTO->~%mzb@s>S$@Q;zD{J@&{!R-({8qi#;qUD<6GvWuaT_zl;Mwc6GBh ziMS-T1)`<8I4oxnElj*BFmd9KMVL zs*u@XLmWwNEW!f~3sK)BsY1#(!BW0D zF>IRzNP@7i$Ty1G87ADVvSLCoZn*@&1Lji&(w%~6H5}Rp<%^mdu;fb&G^;ygVsh#r z>pLt?>PJ3xptA+c!`KP{b+2F0yAIi-2Zs&yev4%Lyr#6L}!!h@Urnl{IE12D#n*6llJOe@sO!Y+` z83B0}jMOeaIomZe>46}pvgaSjv%W5(kH$g7?#r|IBU6|W4_)@>v@*@+I9No#Zkt}Y zm^OIcAe%QezQl~HGM*!W>%@SfA=hRueI#wta-fT#{25d zsFK_!J$4K&-{qMX8#gEvb--<0h!~Z$-a|7DztOF*ICmTwDM1(--=M`d?xu%OK)KPU z?1S_XWRmaXK`*SKSf&Y}QL4dJ+Y^ljU%@%9G*tFaycdW4*CcI}zzpF8(8yYj^}h9c zL>(HwCVPuRP0T_FPLwwomDyR~}GR(CBmmX_koHK13qWF#Nx z;c_gRy`t_eDy|2>UJGq1UC4ac=PELQnkRDi}ZDsV^)Sr$3 zOW~85^;M$2f1xh?pU`V5{QlKpOWOU|r|WO*U|M5T%M_@w0%^9BX%clax=V@Au?k68KvMZ+gUMh{}3dCP;mJ zzP*}FG*luuoAP3WC*TggN#Z1Kcz*p!1&Q6;d@>2gcYXb6hM@Tz_f(7#}W(_)-W{nsuG zVZ!aL-o0SyADm-&Ib#^r?%Ml7yoDn5=E#+M?RF}Uvi0!}jAT3WR{IFRX8EWIp0~&eqTg6+gl9vCuflRY)WPi%~U=Ef)Z3c%9ELlb6VAVVARkgUs?OaIPZ zPk7&5!CcZbO7C6Fvjs2P(umvSPmpYM;wy#~pz4JueBdtYi={}iH)2(cA+kxQTLEhD ze0Uz%4%Z=tQkD_iC&LtMgYW%fX!<2KLnv-WJk5C!q*RALW0YVA954C+6DAl~%CmlU zE*Ep*&$9(N%sTEcr?Q}Xd84!3ufK7F7^wSY+34}O(;OZd<&Fhnm3qde%#IDNL+ZU# zDOV(nyygG?1D%%p1;$mu3J_cVM6e>6$qHN22zZN|8*Y6^R9p|UeWKWFcV~4s{c1T) zfj6`s)`*b7M>fnFZScxKYaJ_L9YjNw)m`Az`MA2wdmPfuT3ORg9UA_5hg8qs(i5XK zm+*8UJD>Yj&YRKNt01f|ElmR?8U)B;hyi%NPbb-K9juM*kuLt z8lfhJbEAW!Q;;g*7cH}Q3#&|T@n>zPzRX;tFcz}G^oTrT%TYxag_}iWSTP+dYPDce zieJOu{#9i^kQ)Vm+P9EBI6AXD1`YY=?A!?O?{#KgYb$mCz07x?Bq7Ituay{Hg(?2` zGW85O;B>|Na+-hjW8=T}V+*7pLRa}$Pt>9RwI@*jzc)OK5q4&)D#N&E<|oyx=}r9I zZ`6#1Dx&H@aYp2|Hp?fh5ehz2P;WUB+n2OG9U`FIMwq;M-h0(lASRzc{OJ(1r00$G z15zv42V;eEwJ|PZTbO>V5wIJd1Gkw}?}wF_iuiUo4n)mJb)a zlQv!TdegsLh*|ku(jbq6E*7!Lv9*GRW0A9s1I9thXcAO{-S~-_c`CW%P;d_GnqX;I zZYy5(fw2*gS~dljO}7oMn~6z{j{BwR5lag5ocO@CunskAG`{E8Rx5oODj*c9Th)EX zz&uc_NV;QD7F}r7jr7i0c%JhqZ-oGyniG#$0i(@nx&D3rIA|P{fjth2C4Eh-n+?pM zeie<3>id4es6@3TuNUf4D58o>w^r`f_dZAnE(AB=b~S^&d&B<_S4d~%)z@?lfiEtd zQ59OHKxaI+fIcSNt}lKu#E*}qi+f1G#dwC|kwp}rj&vUHy;Bhhur_O6$Bq%Wc)URO z1heZr7G1mM^&``D!f5loPYPH?{`AvpaemX$JhZ)EaLoKUnJy+-sHGeDmOP^rD1(Z? z_-v6s0u>`T=9Oq_JuwydeZLn)evwLcJn8@B*hhwMwHH+@l};G6ON5tiMB&!f`%h#3 z8JYin&&(H@>YX+s=El`<#g7C}Km6#ukfL)>5*}Rkn2aU+AT_*qV0k_~vio{{?-A;P zAtfnOA$BLXb5I!uhIF{pbAJ-xhKCNNw>LF2CUiCqDonaHtK)@8uvLIEODBiz6*K=1 zO>D40-U>HW&+e%uB2oACzY`&}ib|<4(hd9%QdsT@fQVj7){4>6Zo=!T@KXoQ;i{sI z0R34?AhwP_5?4RnbARF$N+5capsH4eS~$~*9_r>40tqM#oW8FVgpn2=Aw7+iYCd9Z z6NKSn+oa@ir=cjURiF(qKj%ltF?4&iXCg(uUhSl1nUTw(tT{cxXYY;Q zmCFxyQBl(-uCaB(S6A*oIPzSGo$iBZ{}$B#S#BNVwUHpSqTD@J{!omSDe!nFWDUn$ z%p>hBb<$$`wkKv!u-GxlWG4*eK^@TMwVEbgv8`Ta9~0W~_fBo`G$6_Y$ig@4r+N{b ze@nepASVwKZOxvYWqk6;PJ;IF$&ZQKNMQ(vP6{xQ7tx5=R83?27{X5fbJ=wb7#8W4 z#s9IQuY6&P1tT-+}T2nPu7qdY?H!D$<=^MSSZ*`kYVSLcGxR#2>1Ea}2rjRZgm z{^nFWax^9&y=QQP3#?8p|Z-n2>@ila17tHxKoE>jHtO^7_8m|Qka5m9YD zZ=1C;HB=dO%E8Ef66P)>2uqMm0PLbuG8bm=HKaqVJ@v=m9;V(xC|d+zX#HsPUo_bV z3v^>ncUO=U&`3#SGKDN~YvYa6%)AjTCLb&a%kI`w-L{X7pnYLjhSRG-M?x>C2Gg!j1bvpw*VShBfOZR1s0=0kYwwEQ^`z|sLDTc=$Ni^!3M~IBzj37@Q+%1Qy zg-p4b3{lYwP-1#L4k8jo!8!Nfk$9!9hfA|J4qC%~zGE@LC5I~Qgth^jYQTy_dAPG0 ziFNLQLzl?DW1-1&NhrWgUjvAxYOVo1AudGDN3$v&LR6TS10-QbF;$na_J}X!c``=@ z?8g0oii#)xz;cUWbLzd3KqRE%aSt?CoGGrP%(Ou)3S!t(8Q@wA8=fFUdfpY`NFI)E z8^bd`GwDnBO=Q+j@WTXSJhb9S*iVNBd&s%U89YTH!8?htx3I)D7C8txvdLtM+<%!H zBOxj}u}g)q+4ST)7JVgUbCE4Zg?MQQ5JaDHr0eH}iV-;%nQ(k?yyKxX6&dzEA5?DJ%8Lmcc z#m2V+|5Mg^-TzQl012NT9RLDbfN5mH2$LH{`(%Z;vG9EcUf1Jr#A~m)e#|*3ZkgKX z#tnUUzIfO6c|#5{8=+;Q2TjC>#I?NI;3!GRqXdovGQ?Vtee!6=LES8d^xs6z5qnYW z8o<3!V+Ga2m=%is+`mzYH1HudhyF9qOv5O_b>0!T1SdVqFKe3xf7~l8^Ju*{eK%uM z@O)axoytIuPsF=7>^14Jy?2JZr9TRPc*Dz9)?hdryPc_r`=J_Hscuvh=kotty|R3BKU(;(EA! zec*K|l&W}Ah)%A?lJ2a&HO&UTQO^Np5EC(FFf29GIxBl2+t&c%Rg-g^f`n42U) z9@2RRbv9}#MCG8g_I`P1iGE)uB7tQC$O2r$;JHlf!&WNk(3W-zd(;(qB|SgcV2+{@ zQ+HqWkX$nsnc|N`h#O+~N?bHuj-=DOSL6r{=<4XWRj^)Ul7S@Hm{_tRAVMMW2Dnw= z)l;?JOxaJiscq%duzAW>j;vJ!j3+ryK1V>OS|{44r5Bd44@eRs_@j)q*}w#@OpkPv zNq!=9aX~nz);M4?neS9w|0$l{eL5?c7xS!d@hD@R0GylCyfFK>5nPmhi*gI=2tZ3Z zzsjgWN>pYx>}B+LqAH8>W}fZd?%BXOE2V80Cm8h_4$cT zJ*;#*r4bvf?P=IQM{7$FIK2*tzE($t1eU0wFwn2=H!yPzP!!~AQDqSya)OM(pZ35>F zB<5d&LDk?9qvl%~b9En+m&>RHg$_(v{Qkwu#p;x;7#-B6^+Mx$Z#&k;UzY5zsgF~m z)od$cz<*^L2V8YR(?W$6YctGQyKG~Lx=b?s7M2#_NN@L#syqbf{MxkNKe_d?(Nx?^ zBC-~sVPC(&$2-9+d@lz@CE6v1Ea*N!w*Qb%2@H!ww2Z=ZSgqd`lX;#Y20Z{xN8j=e zOEFKq20*7zXgPlncZrcwk1!C1;Yh?~fZ4`=?mgl@up28c zWfNq!szup(oV&BHsito2U632w{-xYFNN*-Jdu5m1ny=X@^GA9!EBjS@TsDr|+cGgW z-nfwvcV+cL32y=N=5i#Y_P=`1i|!sV}wmD}f0#Y^T^9T)v`~Z|6ULuu`L6(aj}kX&m}k zN?sHvnvX}3quMt72{&E!L>5P}{8wY@71g0UC|efVSkrR(pCe8k+K^}mElx_YnAp-+ z(D{n&5VMXTS~l=Dg%GdjtG@ukh_h4tHGnwMiCsrkVcgcUs?}Rqk^Z}yIbZJ>voJdQ zN2AdKQgs;9W5@7!3>u333mmYmf+1tH{k0MSl&n*+eznFJw$w~bYTtT6I1<**D!F`j zB*M-cbLo(6XyU7=Nv!uepIsQv{TQo`kf|Pqgf(CA?<6so7?=Ukj*Id??Md{;=n>lh z{YYYSrpK0d4~T0(ysyHre$Uf&J34?&?>jX>VXc5_e0}+7w03JX(^XF2cgX=qqUMy{ z(%)#1W1SUnw8UI8YK6V-K0ude(~Ow_iuT?TMdj2X{`Q`>oDgaMyE7uRzu>wF#`MzJ z)`CcmhN;agginvSgK+#S&gK1aLXdhA&AYbWBmFh403UY--XvO1jqh=>_v*i1f+ z`8MK3t~;~AtHETs_Pe|U03w25U5};jZ7lI1!$F7~F-lm6iCY4BIAX@!%})o;5L7N? z*-bPD503_4<#y(li_~E#x@=Ff9L~&7v(J*bc+AzT-{3Vxr!Jdvz!IM&sVh9hbYMT@ zXX?8aUczK1R_1M-H4=5n22wjvKXLbYhbR{VQ=FgLCL72jgpLK)wnhgh>naew5m^$d zrmnWf18Z+;D2%w(Ho<0TSrRvA2yI{=zEQG{gW{+VHbp#n#TnJHht4%`MaMyj!gy>S z+b5wWJ_#a-iW49ot!+{S;OzDY+n^{Kks1qcQ)Pqmya4>sq)7V8j^w$=gy4)gA6`TI z&T1RgOvS?@d*u%A7Bli~caxQ$RTLqY7Huj7(R;k|QkGxDW^I3N+{nnK-g%WaSuyIXjoHUl<#m5$#d-##M2>J z-Pyq0Om)0H;a@jj^f0_@yazN8q?rnjp5lYu9hN;~sS%Jz9|`k*S>Jq>8gcn? zDB_aBueBV1mU25jhk2|KhJ-*KHYx(BOanJ25tw23Ti{20oD$?Euu0b?PJbEZ=kH2x z8ZDPfzYEvebp*GbE3S46=(rUwDoZxTAz=Pl@@WMY0u*Nkn!eqmp+tX=C$y}p^5`$k z-?4W%u%}5&O8Wv27j+K(?cXV~h}Bl-7op@o)Vf99o@RfT9YI2WizcN6f55!cZrtZ^ ztHLF@rbJ&`l5>LJlFXivNAY~e%jpjVAO0JCM!@o*W*Vj^09NL-Axa}R`r7xL8C$OC zZ_MZ(ZYujYq1w0#dxdNeJi??%Fy4RGUY0q(t9+)di(tyicnqQ))2<;g_y_f@a3n$q z25SOq(EF9A4VUe2P7RT*p#8-&jCnF}9Af^7x7ZR|x+;qLw zov1S<<13|Y)jpNLP35Lbx@ZSsLBwqG3U3@F$DFE_$UA3JK+_I4e1`TqUbjzW!d=8< zwJ1XJAr%lQ_Vj;OzE!eTsX-d!{;SU; z7bM)?w5H-^%mww)NXirGufbgQBH%xF`3>>ax9vl`yAy^(0UIsnUv8Ds6 z7Gi&jmCi1NFxG4lgiPWh)WM!qjWH2ta8u%yguU^L*YD=W?+6e(dTs1@_Iub_58HZQ z+|YC5Le=}^{enJE30}_4M9fu=$CX*Ei8cN`Mw?$HCDpV==rH$knIKTvXkgwoh#$1m z1zVaRENJ0H*z{xWbkt40mK{aC7~RT_b2nMi`dPLoaAM;fkEJf=P@&PbMD3io-~KsF z9-2WUAC<(IMoax%a@241i_gZ30-9J5k`iJc`8Hm}m04bU@XME(DANjDgmItb1Z|%Z zh#WN&4b8>P=gLgddnY7eSu7>T0Y|}(N`&UJPLZk){o%)Sg=vVCy~;4nR0LGs{6m}Z zzl?yw9$T@%*+&ov+UpOS#@O4kkSnTx15DdO0Bi9-BpcQDjK5Il+b~G&41u@v{%~(( zNeG1wr>lAbut^y=_$zEO5uf|*`>q{9S5j`NPQxSD`1$XOr~AEJaJhuq_3v}e@g{Ab z8Q<$r()I&YJ4cmwGxA0NQLj!Nm{7eHq0ipNzJp+_IXYXxS}w2+vNy=*8P?VMlfBE& z$~~Mdi?Sd?zwvT55Er>83PQ1kFv%g{%%$_#6flzv7gr_nmO#EgzRCb!sbXj1b3-Ev z!DR!}Tt#A?JAy|ei~rG)^WBVr5C8D!WCV{!n=3=4wYqaE(=YvtN6#&|n8Ce*OdT=j zG3@Z&?wPUlFU1pcDgUG$yNdtdTm-yfg#BL)nE!W-e4V}@WNez@B=-35(;S28^g^PG z@#+Y9u5zR=#iLlure#`F3(GxEr!S*)Jcfvee8I6npLmnNY5;0utmvbWzm zkPvi@8o%Qh<-KRHPr6<0;Sx={pwI77IR0Y4%>x5RfBTIr(*u0&miswg=JQbX|oDwHShnm__^dlcg*7BYI+9gW<1RDYE^mYcenci^z?Y54wsgoR zT}kEN9rP{@|70{#R?OCKjsP3tl8BH=%8`CpuckgBe{;urFyHF2(P$ap@e!V?YYHgLT~X(j?FTZ$aU8;) zkZ>n#fx5TlMaftM_-Mo4o#DNkc@D^PUKZPbY>Dr+c7u+dlDo#qh>u~(!F!4!xm9qA z%39^NDgZ+wFzSO&-TTl_6BZGnyee(|)4SVf2*wH0^PK{=4B(-CtV*sx5j(x7yvYfw z+fl^I&D%)X=6T`ipV1`R(hanJv${L)Z&g}eU?j{O?5cI$9y)YwK3eAu3G#n*Y57uY zzfI@>yQCRG>C=Gpgb|$T$fA4lP}XFpjY^(`86gCgP5=K*DO9l2=01&jx>H=`>wG9Oy*nYJno{L%Y~Y-yj5BOmJ26ga>+ zds^tV{8?IczCa|V=8!s<`(VKD;Ej7)ZS_r2SrdW3E9}OwC<+{`Q~!mh)g>1@cdgR+H#EgGb%Pz85>}T3wp@+qu)i}^Gz1t`RcB@;&(EEi1_hUYy z;epRQw=inicfJ0&@JI`9-Gdwur@2s&0UzK^JDG#@hl9lq(eNWA1fEjrGa^}ux|XNz>z)qxd7i_;YmCE}V# zpf6l~GpzPxA?sH_Qve_3DF@s|TmUA`AP84{q56s=K2rjiI<@LXz5@);K6Q{ra3jh# z_&SWgY$!7<2Pmrh$k7NmGv76>$l8Y~7e@jB;(_28;`tu(+)eH!&}D@8Oj*-wRto$& z9RKRVYHisKZI<=(^EJFzRPzUFchUjS)!M;)g`>9<-zHKPx?q0Y@_BX|o!On98G@jR zY9F?UPjW2=+J1M=Dj)p%Y{-A1}!71jp zZfXCPp=Rjv&KijTmFrzBh!|_O;m#lCDD4O5Wf=y=Y4fU4q^4eEDR;G^VRuV{IsuOox+tjrjR~bLhU!oo<+}7v;#Z6X&LC)zy z-%F3(0@LG8iH*zE!CUP?i)5tpshLs0-JHW66MO_tOLFz=P9?OCI}N@C+(Cm(O#w3q zLM%_t>q6Ys8Gby*Wvcx8knE9V+@r@ykdHBnB4G>u<2>*!`&g(mI(ef|rD3VX^rx?2 zT>{+cA!kWAC3caS5l}$w?i>^0u=0*1=!JQ>lsg9;4I>;)aKV!}S3cten+Ro-Ko!A- z(2+^($j*~@bI1+x-qpud6)o5f7BHvZCpafav4lAhp=sQE;1~l%Vs6%nW{MR)AOJ6d z12@4q|7RRn#ix=1;q9a_hX1|+CdtYh>tTypolgd)kh)A}RfvIUGL?f*HoELiY`VQ9q*}_5Q{6_5I zGiEpyJX4;HEr|VcCAKDhzFp|WB>g4?K@&DLrp5tyI2qYMa#^@1d{g?X@MfJ=5%H$S z`eyRwRA9Nophe$fHU!|Edhulicz&)jfw8ovNA?^{En`Oz3XAvnrP$43Ja56KrPN2A z^-k;l2!$5G6l*NCgUTeIJ5_!Y#<{j)CZ-%QWC*U2xo#-m#2!xW1i~}88tK??z@>HJ zfuAv&F<2-)WS%AK2o-Z*zP#I>Gqk~5@L0e##LAkxgml9CEr<}u&oB}In<$%8-@a@- zW*tKKG`_Kd>61b|&6{K6@S<{BJK-nB8WDb3VV5?FgB){NF@Ob5R(ky|k0Md920%s=v6T;k|1und zdWZ|aad2yZyt%yZZYabQZqnI&hv&3P6a}2>D`g{j?vEPZl<-PoL|->4jTFASYxuBW zadKNP zyRU3iKs^7Ax&$39Ork`}yU@vYA>b{)h5BR-09$G8-Y0aB;W5exa!WkoA;?Ld0zFF3 zNAhyZ0it2f1MH%8P`50hUhtAgPgw!@EE-daNL4aLV!?_-+`L+;*$ZFpbB}>Y2t~gY z%nW)Xoayjv+ETYD)9uBi?-C*#v-*-~ac;J5tkmzITk{14&!QWau@C(` zXe;&gSnldY`PrcZ;3M>M`G>)Lory!!^32aDFpkTwKvQ_8NvP=$&x`48JdXgI_KmZ# zvHgkybJ!m6&Z_+~(?rPA1hKp=m>J66Oaq6;-UUoJi-QmOGeFm`L7c}oxBiHsJMP#{ zl5*!~fUc=+?`sD?hrq5r8y?I6Zq*-%cUTkMg%?f_nB)>_j{%- z82;}g*Zr#GPv0LSPC^WJTO*#&fCS%$KE#Bz+Fh0Xc@OKN)-pwHR zXkxT_@BRFcn|E?znmZiqj3@B&@WpG1SuQ&fch|x>wXOO~H*dNOLn5=e0?#|Lv5!0+ z>HN^EWe-f3Tb}1W&v-3f-z)u@~`_jG86s3>%aK! zC#G=S#{f&|7Bb1?H!(2y`8n9Q)pb69?Hdq|k!}G2hPJ%CR;b8aK3-uI^K|FG>RJsM zW1&BdJ?o`_nJ+eb=`F&M5eZ+87oE}DkBWf;Jkji4Qd^ zp>~1uO-U)a>{*%op|JDq9)b|h#o)S-Q6LZo8v7A^Y*yQGwB#Lm!Ef@p^W{-f0hj+u zVAhIbkXHL%ug6##$8IKv^nToS&zWQbj+fd;&$~pD^+Ox7qQEwo7FR83@tOse;qydX zCjB*nylFdGsttctwTl)}0-Y3mycxQ#9_E|AZuk%9QZhL z^uZCJ9sZ2lQKK1TQ`dne(*{T54v#I|@Xph(PiN++7Cb-!zS%a~I|jCWE?V*AfDKCJ z4TFNauci>^M(J3L4UBzxWYkMw-#ntPStJbvx4L+c$j}lE1~|wl_2yV(#zN<(y^e>; z&KuG`m7<9+@Ldg+oOGj({U*}o03a6NaK3;U2 zd*29$-DV}yh-EHsjifsOc4Hf@2jab#oE*}#K5vz6lN8)2LcR^j8knDs8fLxoUn!%# z`&2UK4+{2@xa#y76C9#7;k$!fjaiJQ(TjZ#urUq>f7pgyt%7JhyHQDv0++V1sjlK1 zzw&itW_2CfZG(41N@y9yp1iQ=QZa6E76a*YTXe5-mRwNpd7V$VjGn@MU$$@EW~G;G z?`N2Jm>B-JN*>DiD(fEv*#w08aWx5V(;Wv?$PAqA0)IJdUwqZ#smOo%L;_N(H(FJZ z|5O%IrM-M(A$8@l&2N6b)TrxT({=6vudV5F`@mng`=JdMpoHUf*Z$c&g4&WfX@KGM zDwYu*$?-Y`J_0z?Rv&np2#Q*P*TWhR;sY;%=o1{l>z6&H$#2dX&g<63rx>&zA@l<6 zZ6wSdXh};>0Wh6F{qmfk4gdjtgBG-?TNaQ&`(EAMqLOK9F580)MD6y+ezBN6hGXD5 zC!J0pycn`;*Q@obx!jE_FZr*3JvdYB^V^0@0Yewv2fQH@2~ykp@g2~+hbK23KO0Sq zKU$f!hxA1^YPm;tZqfnulBCauy=iMCh$nO_!Y-0X-jez6Bdmk)pyf!!zW~-Q(G%ksq%6GAg#BX<=O}l!*VM-H{ z^CTMLUT}S65rl4)M?Vf4#76BSGOj!ZP$4F9?ZLNibAW?q$JKJr41_CE({z@PBI#V1bsQo-jPh!Rg zJacY=R@EfGyk=3d>$1M7Xr8Y%vB-zNTIH4+!=nkt6N>3A)owd%JfGZ#-vNNkz~|ru z@7zxMkQuuSY2@>Sz5?A{LKTVn&pdOM@r}K5SLZ0}QNv0EVZ0gFX&X;q!Z5bO9C*qdEzbNZEqM-Id>VL8@~}qE8(uB0?@6lhpdUXJ zwzkXvnlBVK8!#MKJ{Y5{^@zp0#zd>u5M19R`NubIe$5i$qq7YwF^CSHhpEFf40s)= zOZDOx6GC+s6y9>Te=nxAy$!4sm}7~bbwaoRyoA>A18 zmd+!;IvxwyS=P@x!1Y?hn5Ni@kvTMmpU=?ZF2cJvjw7=G4v9%$g>Qus;4j;u za)a<*76yI3G=?z57xw#>{cyQ=9F*{OfWpi)^vWOjQTi$nWE^@90U7^*!>+PiYCDwO z^ABAQzr6+5C0$MxFJ-Sg6+wpXUyM^BFPJzwHOwP67%TL zA-A^aa2iF^na#nE1MN==&#pd8rgo{5oBifdxk4c{eG+XES80 ziZ9GSp$f+B-euBNGxIa;e$gFUVn!G68-;@#k+{!*5FxwUjOFEkV7x$Y;XS9Dn+c_t zu7`VQ*-R@^@)bjZl6hd7Hk7^;VAuq#Oz4zCTp3Fdrm3VXNeLuU-n_U!>-DDd>G|ufsUzY-Ge!FkhAQvf zH(ni$AFR$QUZ zeDEs9k;r6Do;sKDwoeJ+MufgmNJA&XzVe0X?mIS&qWWdZODe#goA3nU*w$8bpHK%; zs<|2Q_A)fk!oq`d?<=)bO*##E5bi3drcUhf#YwAzO3meiCpKb_XTWp06fam5F{a3X zIUd`;yjc`HUJkz$7H;~L{-lFwX5P>ju`zUq!5mt)M913bwyMaf?2qxXB)T+0{1Pcy zj#ufnI+3o2RsyU{LCK((MX{h%LL$SeujjxoR2MgD8~p4{FvM#$Xd=ILNyA##VMAxQ zU<(GT8lY|{LXoHA0|W8npb&*=D%Rb8Ix05&Z}wL{d*U>1vUI(RZ*Cco-z!ugADs}x z$Fl}y!zYkYChmJJa!5Ftw1v>dIU4VGc z#%mqmHpVP?@R>zyJmi7Co^^X%DDOZ6zS|w2m6oqR4Zh#BWfjIt%{l&QldK$pn2wHC zMlkr3q|AbUB{=8WL3|jXgm9MU9QT0TpUbhkGyZgGsE2(8P`|nnhW{2A1fs`dMGJ*e z>BM$V6~vwkLllwb9=X}mNclf>eQ1m|t9_ZCMKC; zC)VZ^v$=aOxE+7cF*8|I!vqEPtf)=kV4p4N?LJIbJ6yZ-5rFp?q~{<-L({esc9F$RVw zNn)svw@YJz4t;yylYGT()q0q%Cl&`kAW2nCzuQQZ1TJx)bP%#Hta|aCgyCh-$3+jV zWR9Z;z?-d3yd3Ik4ZS$ve~O`TRn5ug5b%iS0p4Fag!2?Wyf2&$y$87zQyAV{DDr4T z)3_xE2_M>GpKoCw;Hs|@|JU?@qbYNjkvv^!dL4JpB?2XC&RoIhItkPVe1Abh%DFRQ9 zk2h9oq8YGolx)oFg>+rPhWfq+YtCU?2XfFLrq?f;;`2nDEG=p?#FqGs9ULO{JqM7V zc2lEq?>*~%Dr76Q81pc2(fPKFc?!R*C9(f%Oy(&`FB~y!C z;Bfs}`iMDv)#&4W2`fsiMo{patf>0ZYi^Z2nlC5x+D6xxGgEq^>w~O+NIBQy+!=}% z^gu~M-0#*_U8t*9;%S=8nJ(Sf+Bw0S@Im^ZvOPtd3#5n!AkXe+Lpg40Wu%CaQ&{l@hYHSTEe<&Q+w6=jtmmNLKN^!z?VJN$`&VHFpf;Z0Oru7A9OVgz#|ox zuY!`#|1L?pa@q-%7Bau@;M&FGcUK!qp@2B(UI@9S0+-Nm=VD%HeVW4kB6y#_)9FcD zvxu$so9_uzy!Br?YRI;?Csi!qR*#qAgE&2+#D5<|ad{2Nmj%5@EOY$tV!CPyWLgkb z);T92|L+C30)%seO1P)gRn?~_1)dG@oRFxfw>e*1uqGrD#W(xv&4NTh%em0G+x@}0 z)jiGZj^82^qz5k4Zepr>M6-rg>%q4=>x46UXn__{9ea_DaN!!@yCQj2Wsu?UIdpbs z01{ivIa@(A`PT?`BdAx?d}a@l4w82I@sngviGqxCP)`8$IJ#F{@&2O*pdTl>-xoLz zI^BD;bqE*2{vQzC^Z|jai-3X*wrHB@d@V`qTqTnKHN&rQwo6^O*q!=1;^wchBv$mrlHYX2DkF%536xcIr=k7$Xv=R5+{-G?EV zHQFoD$337i9rVU9{D8#J|A2_R8#p+pZe+Ql=#05hG%oK+D(SK?bjSK*Z1M9$g^~;f zp7M!oFwgFWw)!Hy+nv9HA%B{?F9%HTc7Gy}P~(?;J_zt}$8YTOEHl$sE&z z1S?Gdlhi>!@iUe3pYdo2F3syuM`oj`=J;I;h?Y?fgtOyu?!ha~KzS$hS0!Z9#A^SH z?WUl@VH5?r(?8^p8t2#`2{2QXE2Kav$yM^};*-oRM5qKsMpEelk-7y z@*u6-Ixik}9%zPt{E0~5+^#Hq;pr^V zgCG0Z8+37$^aWx24wDEUg#j-8V2`hV-CiBtNIc`BbYpBNdJE@$<6qgTx$FG9$6ujo z_DLm>Pzi0p4}?2D@bPJUeQJd##Dl?85SUs3YSF9*pU(Vcb(cp-XvtI7upR0K|JRv zbPf!mt0p@0{|i6o+({mJB(p>D$u#S(mExE3DSsZv-OhEY8U@Z*&rK$*RYP$m@Q8ZO zC%z9QRsSuk*L@PMFFtZ2X_EQ^FQOuKc7{=nQ{NwF)YUZQ+(;|?u3+nY>QjWcnp5;i zyESK<(`RZ0VH3q+{9vxNFc%jgg-D#OjK5}FVeumWs8}G~(r4bkeZu6^<}a6sw!8&E zLm z&p@hqxoS9g&(N*qI1{ru98`zjodrrPJjZW2;4mmuZDp|Q1>mu)Xx-|sI;?2YK+Yo2 z_#t%Wyz#E_9u$5Ka;FIR(P{34uX2}ldPE1j)v2PAj0PI_&Vh|84*IMY9(m*E-v{RF zso>_y_?Ai#1Oqa4z|V(>l(n~iz3W6x@qNypc8zDcw`b1>(COPGq4E7;meA$lvtg`y z?>mu|%65=2iSfGJ3LOD^&IzwIZ1K|K%uv?5@o;~DXlUa=bPCPnU4vb}ewcKoCTdPkp71ZG|C*xUMQDO=TWHgMe8llbY=u53|D&9fFrFrQp79wN*% zqEMAVs?+{imQ7v9e1jXi?b9cZ*1ZWVxVp&&li(!Z`PyGUESoCWu>~Xu=}8N7H#p%0 z)BQO~bU2WBh>(I3CK26oXkoPcUa-UQLK$T;Is7oMss|)d91yP%?!bh525?ph-1D9D z48--D5KDtl2y4M4!WZj%CKe&8xEwfP7Sqrf(rDqcd5Sm{YpTihUt?PRmW4okzD@Yr z_KuaP7*e?3(`;e=%~5H@GZ`zV+?Ixmhjg(Wf{5Tk*YB7dq-r^b|7`vxfxH1c9u9?u zKT-m_90Y!^(6wXZ$md$sWW;WuZ>jA$KJv?R=K6;&^m*}?A%welHMvwxS}VM z73|}>Yj7|daVlpR+I1@{t9t0s7~=%IutCW!a_;TnC(Md88i0kok zAtgDiWNiG3yfP9r0R#pR{tk#&zdwfso}JeV!9P?&kdQ;9!4Py60-n$242vtLsWX5$ zkp%z?8^C}7)-Ux~+^}0D^W|N(^RJzvi!qTdLoVe$=ej)tPWdE|b0&Y{Fy>FG&zZo{ zcS?mD+uDaUL`RgKE&=5XR5h>iiHQtnm5Nm~o&3gB__ma>`~M?HP|D2d4*}gW{WBnI zhHPHU@iJVRIJ!Lpze~Nc82l|*3k@uztt=kQ{VL2QeDUsgR5<*InW%3qx9~@0phgj` z)^4r1D3lLwK9%V)AcJ6Az(!&oY&Pxv-tISqo&Cea=!I0}dTo6}rJ^luW!xt4uXP$% zQ_=mgQm9caF_F@IuQ&cUgtb*B0*0J%M4qQxhjiGWrtFdgOY?lpUevp};xJjmzlgVH z%l8Gm9z+4Q*3b%0o4*YYAb6?7d+z=Op&A}s6j)hDHU{T!QLj{%++Kri;vBIW`tp4c zr1|rIKUliIGS}Atqn>K=nBhC|DzH=J59npBYhWk>NafMDkCZ05ydvJZ|3j38j%u(# z_ZN{&_y=6t<({l5l#-g7BPLI2zym;_p$2LkgJo~*3A8kiquF`nscmD9)y1cu5K^E) z-(X?aqs%0;mF4FIj|q9h#;T9bo+(w!=&W@ z?(TE8+uQCRc3R(@R{0QSym}(#R!Ij>(~*k}wHpbMD8y<>hr+pLlZzD+NRU~0?JS(; ze|#$c;0IFQc(?KUT$_0py(V<4dLugIyifU_Zw7fQuDoi} zstj@4b_{uA`i_E-$q(WB{sgw*JQ-w=A#UjtXC?t%51n?%%0qmTcJBXzar|{M@5s4{ zR1dLRebVU#`RG1rIR?AVwultf&WL34{Mz%rFHuA5p^Q>4=GdL%`9)5IUEBW!HCVvA ztVv2Vm1&{^D6)65Yd(0>|B1oULstQ{_Dz> zw@LW__ebXj`k)wyyQI0F2Owx_|GI=f<6^iSx418qSEtAW0;Gd`#)gLS_*5!;hJHqoa<68tc>+)KPY@AYi+x9P(Q70iJuqAJjHSRUm zRr2rD-Q8d!^_)ICpr^|2MeX)?Q|RM)>zb8P77`G6Mt90f5!2 zlh9*32?Kiw;>6>yS^lR5cK~z@-g*bSau)Xf1K9<}_XM1#6~mOS*xptEs);xAf52=1 zT;Tqw58-${-(u`n>%R6gA<$pU%w`G&ySc}l`lrJ&j33E#rKF7~%6KS)#;>gZulU4@ zOuS-3eKvF(d_;-dtEsx}?YN9^aO-mn8w<)gfQxDpF9KP(+VEoHpkej3hriMZ=ev)p z2soxa`kZqQ&X@p=8-#)?82jvj-o6Y(R2_1z@4LKKqX@|9ck4i@fol)O7~-#PLmmHsc^AXV(j za(8q~q-#q^^T9ojU<+1K`@%}d0UDGf^uhx+Kf7Qf( z{wSZLn6}popSdfGyX4n6?qd5XTbnQ_jf)~Hw4B0wD{i39{#e5zJCulmbRv(8?C`R@ zG5jfU+0b;Z7%?z&fCy#|K>mN30}X&ayP1!E!Z~$RvBPj|jIEtMU-Od>ZD-`Xrck!< z|4Sz@W`USOZ$)Kt7}*itJvBSACtL$T*Xs|(em7`A4plMiTuephmG&$YjIr{6+dbl? z4lu>KCYNx$mHRrbCwrIly^rfouKXw}*-3otN5jjmE>q|)BnTH4j2@VhJZP9_=H+)Z zI4(0n?0hd#N|W!#spB~u<)y$h_Is{zcy*%ePyjyw{*eMDhXbbIR3-NKS2%V2%N?}b507_$R}*Ac<91Q+7cHF#v4k`Sb|66fO6;^w!ZvKxp=-9C-RY~sL&b$&ng@$d1vk#G2 z0k^3ss~Bup3uwi`(EN=iwgYvd^y6Emn981TFhuwZMJGO-y{xfA((?&vurCiV!?1ag z#gn%2l2?nu4$swnwHOx-*@K?>*%3Dn7Kx;u#?^EyVxv-rNl#L~(>Hj^oPZ&(!L1;o z19RbbxwX?HFTDSHJyagyxXCR8BmNHqy#&&7NOgl!irb5ZKG=M!Tyl2dJi=qEi|)HE zYuawd)X%lCZ{sf1isPG&L>i$R7_Y*yLxU79}TwzK7Ebo90m)vg&W;MxVn=z+;9f;ZzKs!G%s zl!!{WN4v(~p?pzo_h7f?rgLi8+XKun{~mOePWU!4E*NYYbGL0*f?0pvctws_y`nd}4#P)c0=ao{uOTp_P~N8&83 z?~O0)t*qtcvS_!Wd_3B;NS%W3ita7BIXzYAl2@YOSZ_nSfqLH{EwoA?w?&zBRvczB z?3C$miDX(mOBB}!^@%XhFYaClV%a2cPe$N#%J8G#K;liDc(Gl71hQTc*NI7>Zu5UG zq^$m7VMjaK`f7Ltf%xLtE28BL&bQI<(w)0YtJJSdFB!YN&I;CusNM=OHQ7DnI!j+T2WC6-|Oh5&$ z@__9r@WAcfdj5HBRY2u9>N2&T;&CnnFV=*|V9MQt%Y@T%4CHp3wfv7h-+t~h^~`## z!vbJ)`&}5c3_fgQIOL=n@%?$3S|yOV4-f9#MG09GG#iO$+Xz~9cCO0WbT+q!)}Y<8 z^-LX5vVfW`zk9w@t>@nx&S{Y;CS@kmRu=}%3Idfsan=v3A8$uHX3hJ~4zAUcr*|92 z2EcMHPY?C_p~1jKM-8)!wzvyzsz|>jWSp)}?BD?ft~-*Zf|jaR4q|K6H8B3e~e*1_u)q1i0@DFNu>2_Wp`oXe=1XEXk0D0f#SX0 z7r=Z7=fPJ5Lx!V<(g{G?ImG@Pjt$)1P3;>Wo7<_J^VRa2pw^^o`0lP9OtevJcHnV@ zeF=Ksa(Y4(8H>4>cUFx5F%0SgCV$0j-~BIBG>M#AKvql<-?E%hc~2eR?e6z)^T`>& z8qFV&lxKM=W%bcXI*2D^a3kcAN0s^czkUMJU>-O)O9N5N8e|PJy}5Sf;~U4a0yEX~ z?>ovtVhRQ?JI#Ago6k?5<@9@LDb0u3#t6uaihHZ{PYB{YyyvG(U{hCl1qf0!$6w6) zPV0+ipMRjIz5M&wEl$b#*g*;iei00ffd}ExJYfq+dGg>xjmU-s3%`* z5x6Ujl1KX|6@J+U_SK#HAsm1n>RHlR1g40ovC>k>Kj!^IC1*d@x{agyD@<(Z2iEm_ zb(oWS;2rKW6TBg4GuOL6Mj&5*0R-F*#@!-+&aHpBsLF1xru8pHak|MUyHGZ5gb?bl z0}hDH453xT`Pk4y9=FJWp@B?#-G_zvQ89TnA~5CEuOW~{*lV)(@6)nMB{t`@-Dy}o zh13xT2CBS9H3TwBAz*yNkMAlNeuGscX0(|FwDY>ZyH$BGCCMDlm?e9Bc ziYCgqLj!T2A!nUcLAY42IZt$2j!j@@)Zpy45=`3cB z1o|^(GvrUE<`}rCnfY*jM?thudo%XW^{h$Iss;Sg^IcC}MGmi?T9! z80&(Z-?{_;rz4ZYd|dMM_UDtO_KWr|lZ`Y*lJ4*?Ud+TFWQuyL_;NM5Re~MmGk05- zN%X=xWDj`H4S_BmLZ0@ViGjI|rE5$H^zvrQN~;YVxMa2f5iQE zisi`IMIF8Oo{LSEy_L#6-gkd6>_8xZ+k^kgUw2_^{4x^)-VDGHXwKsz$)+E@^xSg)gGTLfbjXRX zY*M@NJcx4kiM zJAC^qfiw5?^Syd^T*Hwilt78PHVn0DB%I75)DvV@~CL@lJDq#5(P zQWRlAb_uia0OWXa4UeV0HbEtuBV;?;>zKNtt`y?de8B&4y$w?=0BQVz`1NhPizr6eYiO z1N9joqEFHXKSP(Ep^yq^zc9_~@};@QU#k0kTM4fIwkTPjv{#dsKDCo?4+mwksDn2; z9mM=4*ID2v&ZvG6rCQ@)-u-9C+X&)0q<}qOqYwolu0ZExE<44w`RV8 z$;A8p<7QgCgD)rQ7Oa@y~{G}7x|a4YN(qs zl`R;B2%IR5ACgDUVN4q(-fq}Y)%u@+TtwCOURgh=D6-8N%)%346YX>Q(Ef^Io&AnDo&`hp1f>Ry95UenSLq z9cDeGaJ~Z)%hWb zCoL>06Rhd@&W>QnE2Ne=r}ALj+hPcMpqgZ}ZWj%|W3*JJLo;$Ln0JBd{$un1vA5%e zWR-r3KRi3*f$H@tjkfsTa=z)BApgM1cQfDA44GN{z_pb;L;ac#6Ufp@ljDfTaXWJ8 z=FdUfa!66=ew0`y8&^hlRfWR5xX=_8=#S~SA76hmqw+})?;z3g%X5@eFI9j}^loJ! zsFUfxHJhOaJsCsAj}844`(u4Jnw4$>KO0A}t>>BKpVXB0WvyhTLm%zI`p00NexSHT zL4VI17QEWs8@20{Bm>6{qIBE?yk$4j)B2^TvcxvC1~Yw|4>?e1aB3sAt$;Hx2#%YW z$q6*famgq!gp1|lA~5!ieT6-~tHVz$xviTa`=JoSg`+i@jQ_N4^1wbdYS2Bv5C;da zNnYVgOQ?61ftIost$1Xy^|G~ZOF|sIB3Fqz+?@1Ac)rN-Jy!QUrz@25JQ4sr<2>{mOemKj)1`PYY$}L|gZZ zFi{gV`BR4Z=qJerpDR0SK6G1S!<9#u0AQ~j~iLuK9Rm(3cC9d zyf7v7n_C6lvAajUJeO#qYDN zB53!HD>#-SnRiD{4D79~Z^l-v1EcUJY>5#=kl$33CN%H;kaSR+gP=7el@)0o-uyTI zLRZ8Sy~+MZ+r7-~P&jQ7! zwK=eTl}jS5vf;)i_!R18@?lyc&zTej(|(FSZ#jD&5D+xo>0$YZXlHvw9DGV-e#^1F zRT555SJ83o?CD0atpj3tzqWhJKo#7}cLCxhKlpQb#wDsPh*av}Yl)PUwGtyQ3 zF2CLRr-4}=qWp4t?^yr!HD#qnjKdd+<|`Y!seD?DeohP-UJm_zyY)!X8@HAnWYP~+ z;}|m5Vb0e&H8k$-k@uno#SFU$iFvcPP`P-6{r7i`8lqVpPWEPnK=v);O#Hn?Sha9N zf?@yOyrdJx{+nrGm{DYDv~pTkfm4sw7*$E14>KcXwVxy@$={&W;+i!3{kD&HCu27hEf8yKBI1 z=j!L9?7vYMwRoIz_x_(;5_Hx&466HPyn-_PFUx4NzqepKPMmO0c$h2cw$KsXxAz!O zh(MheDE+~7N&(M&rX>@?QN8{8{iR7VI^tD4X9CcRfqa6FHt+-mLi*6KXkrNbNL{0} zy)^VSF_bRu<5ME((b7}i0+uxTz;~6pbC$ShASoOWsd}B}fdpb5Pj^2A)pgjua*9c^ zPXE`@TP#H8Nj=z77olF>bN|nCr{SjZ4$g1yi6x?WDJbeYo*#pQ1O!77O3IOWHnp zH3Lsx=O&VS+Wd;jNl?WIs3aDCeM&Ur@7NXNhROrs9y^+pvF}IZl*ZV%--oCBT+x+m zY|jvStN*^tkC{qnO){C2BoxAJ@sMMmkeIK6tuF8bE^BU==!m6GzywpZ9&yXqci}J~7{i?zeIE-AsD_eQIV41~k8Dxno>z z#>W?zmA*7J>cm7r(r-5Jx?$- z%=P0}`{6LAxD+&Q+qDO>VkdmQbTzP8<%R$^eJ{?3kTMul$og0onRE{7HywG@P z!c&^-z1KKHRB++RydK`grLCR7DqWlOp|4Fs%A*d`X>PZGmwnSQ|j}1SN zc8%?i(k@61z@n?I{T9y-haqC}fx+Lrm!@2h#Nhoz3BOOI;gi|VR?<_aO8YbD-y+0V z->VTN4K(e{B?2Ogccz8<_G721NOZ(?8`fw}9c{{f^(;vrU8@Tc3de+#Qcr}4UDok# zk|z90y5Y4iGJY?SqUh~OO_Bs2&Ga}1zb?bpE>Nh3HF9ibpi8EvIm1} zv*0xxN++=~f8!kKj;`zlprdF`s9A<0xb5S3Qy~&`-2&-j?La0_!-h3pIW65zZ7iF9 z0=LVfr{9fUk>U%5)9*9f)qgzGJ6P~G#quj zqwL>($DPT-6>X7KEDRrP8r=#ts%xAKdT$tYH>;3)1aVw+kCO)l2y*97 zDU-T+vpSyXfRp)8%*W2&KMpDw0ppId=tmq#WGPFCu?zIj>y9alP1OBR13rG%$0?ad z)mF46UACfldmH)+mP7hBc?q7Y9aS+>j+)5U;4j1%$DSn}$2O{!lq796o>fcEWffX` zZkMjxVMX?v-fsSw)z%5iQu4W5Z*w1ez`+pdSqE9Edmo*uLqnGb%*n@JxGAR?5ED?&Yx*I7)@pH+>L9{5Wi3V z%6&_CjbgLF>EC%?oS)u~v*8VIjeiPRcmm;F^Bs)KGi z-H8;?y#1o-6=gG{W`EbC|Bl|Ap)|A~P&HjfrY(FFRy#&8I7SgOI3)@}Kmb2JHS-3&AP?H6Og4kDKbJ3{)w^UbONkKU;i3J zN+#YMTjJE}>u7!%Fe1fj(6Q$r3>|2G^Fj@`5Rt&^N6qu504EN9xE7Yc1TH-Idl`~Q zSuW!FH$YF)`9!#J>G$?=J>zD4yGL$JQ|VJ89ld}x@m|$uBxm9P_gNfPEES@A6fj5A zhkQzW)qPGnOY7Cr$~iaN%XjE~$x}I0f^)l~_}gy&+`PL~JZU4Vgx>sb-?S#P1wIk; zx<4-{EcK^4{=NA~4cXzatB7ir$_>Ik#)qw2XICpvwcfB;8rDesT&r^4ew0RS%@u%^ z85ztVeN$?C3_aPI{UGXrF>P6vHVb2#L3qenXH->W(^;>Rj%r9N(dGhvE zghXi97!}8dgZFmtx=ViM5%x>O9j*T?1tkb#`Vp1=^~hlI6&<&MB&WYh@xhYc9xI1O z^_5arP{|9A4nq*5y~!Mrh5Np(EbtY0%T$< z_8?BcX<{XOJ&HymR$$Jl;l^4A2C{jgZ}EwJ5&QCo4cZij;H$p+)_EWHj8ig<(|jmn zF^|g!UxMWC$%qe*)sgmJlI|UUy3fE!c1b*mE39%LH3oxRoVJ|LRl&+I$84a7b)iQ^ z+R=K+SghZ%0fl`co#K4}dV=#!9iHRC`TpL3<^Rl}qbt5r_aF}pOU+~4=VWsKa*)W| zO1gwA)bKO1BjY@BOlSQ04WH_Z7N2F%luCd#;O%cRMl`-LOXF+4+RcN#DAf^0#3uB@ zd9ixZ?a}faS?Xs<+5Wmrrq5{WIOr{dSeDb0QPPZex^4&go@86EdHpaqo9-{LaN~tU z)Q*^v!&0x_?R*xSs?IET_}9ew$6hl}a?cOQ0L@V@O$I|McX$F3s&~ZH ztNMkce(M@LYx1@3?*Q)BfahVTFGHIz1AN}}h=zhW5qL55Pe!5omC_Bw=U}5_94zXG zv-ynXRnFc!l05h{Vdj7PYE>~CVp7 zhgVG-i{@5&?y97`y2fNUJnoAQ@s-iEtaA#vgB`yT9bLRRVCr{y>Q2#k7)FUFK9?O^ z0nEI;^dS1R*H!Dk$$Oaf9tb7L;3w%cN0`5w=3>FY)c!groXuSYprR+re}dX}_UQ?n z&J+Al@UqHm%0F&2e^;Z443wLqv;P8YN_Zc8H z0$;Jtn|e7D-YWF=Z{&9g3^ct00%6kpq8322*|DTfdHAqfFgtPa& zZ&VZTtQ8XCj*;nrNW8R%w?_o}HVF78Z6sxT9BpHV=6Q%xN1nojvQC)6b44E|l{8l^ zR%$Wu8!%Bp6Zhuii{KDO!saLsY$Ty8I-aJ4weIJ@(yK?S(FMil#=gFWSC83sv)`-F z8jL)-bmNwuP>_$FT<7iUEmeftmZ5kLQ0u{(m)~)w$ztMB76BpVI&3bB1ES0%fGh# zj)rFe*>4YHU;WnTFQN??GGqBXe00<1a$R^E|G=P;dTzP%Pkd!a5N%MwpGDaYh#+qf z+^`E8sZ3H~6e8`Ed%Y9UUwN{u`5R5;D>Hr9o&2?q#c2KFRyE#^l;ETK=x~a}AaVERT4ps}R%?yGUyI%9GYMS9Zzp1CXWz zE^Spj6D37;$^T^m^ueAI?_MtgY3Drx_lkJtp8KTFZVl}R$cq6ymg_?SH; z@y4=@_~~aKB=k4x`Rl`Y5B2bQ7T7h~c{$Iags!0hQq zNn7!2U!ntU+qQIry3J7qs-h`u6QM^7sNx;{RGheZDNw~zaVur|iz*mW=9_+7YZ=FUQKKj#YZI()@nKpffmyuiw z#z!ApRrUv60=xr8*C8G7PB#@`)t{c5$aDLGxZ+eMjTB1f_2BZ&{c-Bf3)QQ{*BU_1 zh%v3e6@Fi2Q3@WCfq`w*HnCi>AbXUX?Pgtd{BrP$coBND}Rq`W1VW@_g)D#n%xpLXjfy4VV zT2H8*qi?DyN+`#+pCwm9&=_!d_Ip##Ge9WreU9rw=0U?*= zgI?>Pzd5m&4Z?ja{;=E)quUeWnlBr<;rc}Vg^8tHIO+KSH4FV*9!rIXsi?Nxt^IJW zh?n6Cs{4c5jku1t)31NrJrg9HIZK@qjU+tHsA2bE%3+~@#QUfU&;lPHvztP0U zxhhBNrrHjfr|;LJ0d@s$+=^fKe8Ni~Jt{q@DvsPPS~)z}d|LgyIG&+9i>1$6Rk_zY1%JbH>5xs~maQal1K;y26w7m#w86_aoIa!j7oG>g2x z^yc<#>>_>jAD#d94?ZAek8C*NDKWCagFGeQ+8kg_c=8)Sb=xF}PbAN})Nprsoc7XILa03U`^VQk^J&A;c!*l5E*zV0v|FvUBp$*R_{lzLfO$XO!{E z%I>7q{m}7N%@1v_=^|%>?PE{(riJ2|#4YSO_@pcIie}BoXojZPbtHknm%`18u z8Y#suG-lM!4miod1G)o37F2Gj#fsz&^eMed?K+!CysX2fbedD|a-ID#JLvsj^L{ma zV;|k=$WhniHvg#2G^)oG58J@r^7T0)VjaV?RQ(l7n9Ne=t!rM#6yMgSd~=%`PqLUM z?QD#gll8VoFSJy>I%fSS=PSDJGg7BK_u*(=`GKe5Kq+)GbS2cCCeR*#zT(&fzfLTn zWOJAv{h}n|O0M^Qt!QR3l;qsutj5V}|4#Rhk5q{we(?u}O84Ctu4w4!D8HM%^xX@e z^T6$DrKqDa8NXg-MK%=~_4Uw-r;6^M-n%f)*1L1AIoAH@W|t68Xs&aamW*f#A*j7* zeDh_^iFgA8o1iqaK|Tq3?Q8d{ok1G{nAf{K$2oiHh{!y=zg{%6uQaj~uurqSgRHmq zf3wZ_hfAGsP(5?dFpd3YjO|JLoPRlSOaBgFOKRQ;tC0PQ_IcnD2G*Ilh8zUJ{Gr}&ePL#!udhebH_p?( z{_6ep+Ve(R#^*cUrCA+3k5>ojZ!*1S;LZ{4&euP`uH<@mS z(CHn|=i+lwq$f#Ti-H@I8MttuS)^+oK`+_U;Uf;$5gfjAT$aDzrW7hiYR;Z6JdX$z zSJ<`N!7>SFi=EB`oYNrv?NWIMYg}O_T0e);^FPh=MQu9fr?|JCPuzFtVT~V3gaX81 z{SI^0oLATRYOJCaW@#rS0#6O4^Dh%mFL%dg2P(+SQyS_bD? z6ch-2vNn|CeztzR{&c@HEXscNWcN+)#zbvX^WmP{Old)kT|7*)Sx*6HWLD=QxCcYc=jms2%FzEcMd&My{Sbr{Lj%c?zy`p8fU^TV^w1pb6$l6|IaAT|e z@L;|E!aJ2Cv^5)^1{sj|K1)b-(L#dlM#03UR=>9`7ZUaq^Ar{}mgs)lSF%0IoU zb9RJNY6U9lZ(Xj7+T4dYV2u1XQCTB%ODlTX+UY*KGaBI9LJi5Gljj2s>}4pd@3$6H z_sE4PvqudBTz6g1Pj-dO3&e+_iB5Pn2@KP8T9usq>tF>sd<;FmW|CD)@z^>{1C0*U z8n%*0RqJv%bPuqUb*&KV`N?VX47)%8$8*OZ}EZS2jJ&%=TaK6 z$!9nAAn4t{1Ny8cA9N<|XSIH-k$*95r3Y>!1)<^rdn(5s#oyZWCf$-b(akGgSg+wi z{}C4M7wJYT9-Z#v99dCF& zRPbY8kH+*!jEKJz*(r%s%x9ebJJ0;j^XmDoT}KFT)x)tEa%NE<+o?a4=CZkf#cXrb z#vxY(UkDNBi0c`YLCoO}?R!pKF)n4LsLx3=2nFWgTZ&Vc@};_=BcFV_wyDj3z9it* zZ6m~5cet#f1RZgp?AN8GLQ>9j6RTj2h2T?przC})h!ljg;*z0pN=TwK(TF*4Ts;3W zz5u=hJ1saE$cEN@3NpSjin+{FqU*B^5DbIg0)o}SDva^N-DU@<@EB|S|h6q@5Kue z6%LC0OL!|;+AE(oWv&k7X^A%PKXDrQS@^HaS`V|@M^2qGUiFgrM9`!zKG#$2*7VIU z=L0NKUsQhC8Wh^93O@UOH620ju$jsSX4|Z5v4`C-d+I0YH*J;*WxQh6 z{KwmQg45OBtW2t2hEG$@WX`evY!`I2I#J28XqiA3v$*%Xg4_QPl_n(t@m?QHIVXC<@Q0lBy|0hT)CO5M zhP}RBWmb*5mGM!E{>$sX-1@?mnTgEiWyZsOA$QxnY+U|8GIleB_?|9ttG>ev2Kpd) z?^Yf%2^Zu)PUF|f^g_seGE752+NVXMPz9Oj9Y3@K6ILH*bkjxQAaetNh!z z{Q>()G!NY>j(dde$+x!)~ z_S>>dhaNO`>@p#_>7s^e-}5~q?FUmScX!e&3mWoeoiiUFjVK#j{pmRTQKU(6cs+Oa zXjH-H6AMtc0%Tg5C zK4o$o|J`ft<8dfDqo1TJmNDZx<u+MDz6Y*8 zDc}4$^EwOJ3>r9I@(J#n1EQ(ezAyDjWu7kB_K2=~w)gu`cRI0k%&+Uf@K+Os>%^7$ zTRlOe8`X?uOT*uxB|D){d~mV$5$gF|?jE<~*6*1N08-AA=A z=jOkct@(C7e0B#Kf_iqCT9_Pf*q9dGYI1o{93trSp!*qJWVfZk|HIu|ezp0uVY|3n zTfAt2;_eg(-eQH~6nFRH5*&&a*W#1{#arCHKnd>dPJjRbLbCIF_TFo~f5ZE6zGNl$ z$~|+LGXel^zoE?W`J3~tqitG$3_4$}S1mdvP06T|!9^(H z$bhUmV}7lb7YG|pZ?TU{>*;*M0V!5x)&R1G~O}|o4dI5*P zhZ&obeXS-v?u>>geA4hUF2?<-jV(s`vY|hVe}%a}f6yC6#uF+yf)*-h*a$WZzCz_D zOjt&DQvG_S&HuJlBz#4G&I&s-YPW9?YTuGX?qKqEeuR8)r>?_?rB;ZI2s_TDule8n z6rF=%-uyNdN8T?PVY@^>)+hh%6gQ)KtE)+;O7?(kG>hKLPjgcPGPN5RA0Rgdptkoy z#z~e<^zXu>D|Tr=6_HpA?=^kIi_Hwer=cQXq)_P#mU5&;WSqsvmb}I8Mfa<>H())udt3(=%1ZSd;4UwIGg|aDbQNgElZt35@m?e^Htyy5wvGpVFFqTKQunx_)3!(TtHR;2jrqd%IX=uZPjR-Ir2P)Kz&u{8qbGGRVhPMV)9fYJLBdteSAE=( z(L{7r(Ig^mXn1$hT|J96IdqY7forWE*Vp*GDu+DZZF)iI`w_$d9pbOZ74@^8MMo9_ zb_fu#KGOV^8jR`UG90)5J(bm$j*VJh83BK8_%|m(X_slA<*)qi8`jX3D;uck!~Mm2 zk)$+(+i!pL>ZJq;vnq6Xcf~u)4?Mbsm{NsXd&PA_pwqynH@ctG@#1xtB?lbx6 zxAvZ=w)>pFWnYpw3;wlGPJvBk(&n`vEyoiRFgOwY2L4aP?d0S%nUBboe!l(IDOHT= zQ-S{h`|-^Haqn*DH+Zj~&KOT_4~kOgx)xbnRb9_BIbKunHSC7h+&F%}Jj*n7@_UDj z#XI%p;!T#gtifjKadYFNFz&#atv)b%`uXs}$o#_jhFb+}g4^bZosagm`{Gs@TyG{Z zkS$EFR5wEJuh!jXoSesJx)E^7ZHNQP2GyS+X#HM33z>)3U#@W+>}p8McIo(2#$A>z zucgKT{F=tdC(y}Gkux(IozK{yvi9Ed{YFH%ZvhN83o5Eq?X3?Ns2xF*}{*x3C0%TW2~o{Q3OrLy?+ z#g+$(UQ3-xi!2%H8S9gLffo#jtOWmt!qNBOPnS{rl^E{d^hMm3%Ak4g+zL24pL%bs z`v-AnXgSwgWWr#3uMpZyrlNiX?oYxyUt!7xl%hV8r287<$p8pylEi2|;{YE4x(ive zQo!xLepJxnA5{urI3caXAg=T>K`!uBXU~h_qDeliZgi`SE!l|j;0|I!j;RS$Y9|Sw z^oOFqNZTkGtQ`&VT|JJB_XpMs>UkAXPNH^DC9K|NmYg!LV1PzFlE1RWdAU__r3LYU zkmc+>5gUSg7ZWB*B38N$K9r4cr>7W2PMUO4F>_1PR=uJw)?^8|CLLd*4rj1#KQPiK zKyAu8aW)v9JE1$OjUnfc=CHM-vr-@Wzpj2n?jdb|^pS7H7ULp$V@=jxgopLsPksS% zU=zw|JZ}%i0hm}I6tRIxs69RViu1NYemKRf*+N=L{_WHeQXE*rC`zp2JBmsz(IQWt z(lTQHF z`q@>N$+fl_EC|Tu$sSTX9Ue^*tw2fD{e1Us>cP@s@eX}!n__er+xjHE^7_QdMTcF3 z7i~;l7wW`Kwo^44E;iQ?eBO*gTj-0iS^o@24@o`lJ$1^yX9ROGcd7Sg zACa|UDq)QEbOqiI)blAyRU8cT!VN1xp@kk;)KL4K)~{essLCcUGN6%3JzC&*hC2%G zE9z=NVZ{}Cr_*A*iYia!?}FZ(de;ComJZ86)FkdtVrf~{^K!aOOtND2}DFt z8`X8(eV)=IqAa>I-&pB4zbD*~eJ*!8tOXOO=}~1;C?$xvd>#hM=Ki|2FCY2% zU7`d=xtuJ%+UVUx8iu-p%KuWH|8>%4pIxKRAZPg}?a*lkR+F~br{{Iik9<)k|dYI9I3g`=rmrRHsrY&3f>2Nvq^DuWhDCQ866Af3MX(R&=O)V z<#vR`s(R>E5Xkf+bAwRB}9q>_t=Ue>(@LYz!+BhBu@m>c#`%72$ryr+-J z`b0i3B;+~h2`gIM*ZSO#Ch&4B{GS38g(6G%V;2LpsKcB6hTN=T)8}?ugcqA>YauL% z?U=`DgT^QwM$Fkht{fJQxo_68!GR0r!)b17G|^C1O;gkqG7G+)pCMAY7Oc_>lY2*` z(#7qV|5K}6Ko6vsh0Mv}NokK|^2TRe?e8Dtxn&m1$H-QSC|tDO+@ZXN92D(b2L1B; zhel`;fkim4x?||P>$!+nXHlmX4db_9ztc41oNwswq$Cwsnp~MyDJZk!gF>B`o5nZc za9_smqxI?gk?sH+j%koV;22ZJy>fH>4&Grfr@qy3m=~EhJx) zR-^gIYtl%dQlI1ad)pD*xIgxxQmcHw;XQqZQ=P*Ld84(N*xkqN5qdJjpZb@%fs0-l zULn8u4md6kLtj%zT~mzmlI?Pi?q`YWCx2@Ld;W|RFTMI2_Owb&C8EIN=)sV{XxEW1F(=r67M|iL*Gnb~8bz?KS%u@cFbbyZYp0e(HJrLUZe&-|?s&JjGV> zJeDHZkvGi zE)01QP3L#v-YrJdk347*A#EI=UL@0ceYFF2ZhcN#-827O1$(;a*r%D`PfX;gSEGhk zQjjEgf0yt{%{8cW?U^cP^NTG|E-S_30}HvVkpBLHdGK0AcNIrUPbq<>c-LvBtTNz^ zG}>2Ld=z&KZ>JEV#-yw@!?kyaku>=!fZ1f-eU|Y064CM6OGginoQs)EMxUrVj&yW$ z#oO`pBzQ)guwW3dtzV<38#gfJDN2A-zMUaQcd^mcVT{ciibO04{9xpC^ckNn2>eM_ zaXyLs8_86D4fKtG9Y*JtnYd1Jm``_UA-$bZl1hppJ*gJF9o zlmr|h1--}#HcXhR0)krb1@i9ea7^0kcr_qR=@r#0VC(WKVv9BBU^WRi6|W& zyTcrt!-@2DbW-W~dlc8Y9)rlKNw_>P2A!G!dAfpAI`BbQz+G*S%;r{Ibe&oB`TVUR z3h$reS=(NQW+Q502@NR&F$8yQH!xqErf`T*~9 zPkMOXr`-OM2nctBa`gja!~W}w!q4(hgZcWEHW572%sG->xdhtLTGx)V+;A4JU4rw!n7ccsT;ipXVXgmGvi_*nx22PD`c(T=FiO=Nrkaya;EUo_$gE}(Ic+fAUw z^q^t$%7~Qt|@)z4Xvq}u4j@H{sN?uPTMT0;WZ{dk4jRRIrW~sUeg2I zqc~cs+fd>dYKIZH8v`egXnV#D}qQulGW=}>0VM{fJ8hS?qW>qo38CiD!KHIbtX zeba?`n3)AjBR&LYyWm=9MysKNc7sYNN%Qzu-z0+0&rH*BB*LHSXTHXIVU^KlJyDC? zPP64(8aY7xXbEOKN?;ww8NQsXnhn>z=b;~244ayo=CWc`$o`nFj;vyk#{~5nEwN;1 zuF587ZR<;yOx?O@_O!RhcWn8o>m9zu>E*&YhkEL3)D`wAmY!EES^elcm=3WuA;zPq z`Lw+Y(vn|%>6@)Ao(vmpKe^>{8;Mxy-kW|0p5s5s2K23Y(CExu^OZWYff_WgOd72{ zsKheEwzL~q}@VSpZd!-8(=jNq8b6{b~XU9dP#Yf92?VW$NxXg=Zr z98~O4r6p%WY3o_SL2V&3@)NSJXAFtPC1QJjmmCbl_=tDh#;6158L0B6Mb{hT44T{5 z^IV&8VIpd&jdBUM@7=nyUGu3WYMJe$^wT4n#~q~2c;1WuMJeh(Ks@*&W1YJfh^$h- zc0SzYfrWbg(M1NAA8e$W-w(V!%)1q%aort6Z#f?FoD?Ta8M)K)R^yVVV>a233X9`B zs2r?#Aw*32+(Up+6zSbkdECwXP>GE^v}zq%rSp5o=dQ|J;M9*^#i{r6r=7ifHGgyP z5yO+k@IXy`>pxxT-eP?nSxQ({XrY=}TVM1M^jJ#Pzwi7RuMa9`#ha+Ub`rb0OYSGS z4i&w9SsE^hy@%Fi&?-gMGQ30VGy^y+_ERoqBYX3CdQl>8WE!c&;dS`iI&`J21dOkKXJpRW5<%!QvM4@l*oO+J`$or+A2Qt2Lpj;-FPois2yh}&) zE`|l61x3qk{hKaTvZ`IceF%^4-yhGks_yl&VBju>47F`ZmeET5bdkm;X1LbX@Y6fn zNIWl);S_o^%Xcm(GKxB;LEX+)N?cNFgKTSpp9j7XpwoBBEolSV(#aPWC7KOx#C+iVxcgmVDe~2E4t2Fwq%Z8KQ(x140+2z z35*e_jzZ{O%|V@xGzp?1QKpd&UP2dZZ7Y-YfZI6+xkpeQZyoiRzYk7gppZ`izWjJ9 zOMJt>ZOOL18C7ecPUR1v6)R(VgTpoxI~^(Xh_Bj)m0qFzDTo`C%@eX#zhYNZhe`NU z@u@K_hWbL-G#Dup!-PpX#;OdG+t2s9*~0ISm)dOJfExZ1n>x7bp$A{Spg(&rGO->O zflUG`G3q|fG+pVV{CyJf*6LSCGj~*8A29o{Rfucb^7Am6f-oz>B!mEE_Lr?ghBn%! zH<39Lv|;w;HXvFOeJ@|<5fMC9*nOppsQS5^_Y$>G`$FBuKpx*5qV#YgDR)FTeva;B z8Q4wztsG#%ABr_Y{JIgC{#__(uFBA8TpE6joFEB)b;4dO(k0io^&e8|~;^BPWUgld7!7*71ZgdOh| z4&db#eZvctBxHK1WR{gQf+>cLR04LNHLf6Bm+}El9ptxk1%5o-0o*6LVf|DH|M|M@ zp%K4OO2}TQ9`f#STI0Aror$CC#`42AOsL?1&8gt3T=ifwFMko=b8!@Q=Nr>;;7%Z}*|GBJ_rqG|l;q5aFU7Mn+A)lMo?(X~Ai7g*AI zt4zjw254D-RmlY=fg+X8-!2>8jlW4_bB)#%fEKd#SQJ%UR3<}g3_W2+|GZCefD zy%at4M#jJA{OP121q~FtJpc@gQ9;#}yE|D{z5jGc81WaCEzHh*9CM3Kvv;IoB8OLjg!s;jeTEnP}{eS9I#;V(}2xrXzk5eXU zHTzJ5Mr-!RF5}NK%Ob%mU)D1@8E_^H&p-c}k0_C9&MZc&!cCCZLDOzfWx~y-Cm>HP-F2o}AnijuB(^|^U?0Rj4ymfUrh2!M zuS_mLTF?2i*T)EXWT)j88g)Z4CUw8TU4)!fU@5*MHB0npR%zr zin^^OH1KHGh4XGHYmx+l5dxNx0puR9D!hPfuJRK%GP?t7i8_oB*W5F@yT|-AU91tGO6TdL9?yA|@0SltZ(NuNDYL!CVR-nY5RLrf_K4;fNM+F#S4hZM&4IPGCzW)4CGpW_* zCP%XLdG@N-ke7DeWen2vtL4)qt<-*wGp&_BQtr}v=_OUmS@bq;g6IQKrea0V3sCDi zVQ~Rp+~o{1?c%?Q91a*+oUi$I@ZE+j^CK;X;7Vud!aJKucn^+>e z?J|eG>QLTaFN>;jNyD+xmO*ozjMeBh`u4qo%K%?`pXyyCNPJraSiWVb z1gbRJdbr1wlstN2Q1D}m{B5V%+x}Tw^x<}eO#p3u zluiXmf*rUQ78W;rNnsi!yMUr`WB#Q~HtKN99wjK?m@7tfCSg$F_FM6%BJaF~_W9SV z`6#)2f#Fz>(8xH)i65{f{L+8>`X;CvX*-^~iK09C*yo7Hk%;wajR)+^P zRapAUaMQkY+cT^5-&tgE>Sk7ZxMH_}jj&DI&aTH9UNw$q=$lk3iOtNlfoDqY^WG50 zj^YonNxf<#Tyv!qWPfUPd9!-35H3l4bYSEfy}0RhEw%A4eco8CS0?$JHKn?pIj0CU z#aCc=(0_P@{&*m%$MK4}i%TC-APq<@_BM9c?ql8`J* zO-pOIWg}-eVpf3a!_h?6?vp1wI`Wm^vpn{P?u>{VtldiKtz{eW<)R)_24HrdWai$J z9ZUE9L)XM5+{LIlXNOD1Fnsg1QkgmAXI|$U!|{f*xIMCF?o6u(GvdNw)Mwd|M}nht zrDT-C9-y%~aKa#opm5WATtBy=CX{A@o{w!(PYga0T^D+`B+!pY(|FdX?YI1%40tfO z8oMSU3$p8m51)b`og>dxRypCob7(SV4{!HA9Gw^Xw3VUa<`+^|pa!GIhc|!$wrBQj z_ojE}A#I3_C!4+OPp*$-^v-51r0>+%o>d+J0`JOOZ$kja-67sUC{ZHJ77n2RZo=d;NsC&)xqCR^{i(@cQqb!SIU3-6}GdZXOxKk zx(j(ZyRk+cI(&(q`{3{TDEn3y;A%)mci`bcOd9`*SEbtG~` zw>hNfWqM3`lQUi?#m{W5OENb#N&JXmMqXU?Y}VTT!#Xx`Rf4(M%9gW?5+tTyQjiRWL0P(sMI2KM7 zv`gaKBuWK0*~Sm)K3-vW$zn;VExFaY-=QBG>kmGCjrEd-az>i6()5&)aqQDRm-lOy zAYw`o~P<(;+nd3vURD5g-kA9VNi>-25V0FE@pOB z$KZf-L*NdB%<$z!$+qjCq6DDpZ4nP?`bOgXSGO|_!8_sr=Z%W2qB=SlVV=kbqx^F) zc!0QC6gN>1w4+A!%2;*+!#*1k&R>LL(FiI7x?W~Uw?1rD=k5|ep5=P`8hzSl^@}!D zCZIu}`i<=Gc(pg~ac8m;bc<4o@e1LU9V%V}9+IEEf9Ds#+1|4Q*xvaYh&x2lY5L#D zGz*kPxV;3fXT*%p?hZ`@MuzX{rIaXVKtO`2FJN)kF)vrEL$s0i8VUlpvZHveyNB4{ zc@t7DnS!%yrVc)Ch0$dVVs09uMwD@m`+NnCp?3qt5ghWQA^8CWCpl_VSDY&)ExBXU znoCalR(*aNraJiW%y^f0mD|X%D@W%aSGkPj3HN)7K>Q<02CCmwd3(s$p$9F9T2Lv8 zy4cQgI*{MvQB5~M`>G~wkFhU!EK=D*#lpGGkY8H1)tk1ZvznN=8fiCD4U!9b{JU_{ zV1HN8X^(S=VO7iRd!PIDf3KcCY_R0N;phKz7D30tlze3jY2W(_KDUSZsim(?AY=pqXkt74w~;rPv}2UlX-PbLB$ z&n55P1cd-|Jc@*fj|5D#vw2kMxnP{;dvxe^%4u)gN%|!*lx1J#pqf1eV8CXfQ%&!a zQ9>T++Tg9Sp7HB?e65JwD|$ld?XIiG)Am1QVccTxJii$){r;;wfN0pt<9}pI{*IuN z3hblwOeQy9CZd(FOybXVYK9#L+Jm65b>vS_>gArYsk0H>SmMXXl?Umf1LP=P|H1*S}g( zEWV@*(r1zi5)FwkTIF-IA7vZIuwzQlEoTOO;WH2=d}ko}sz1!~b3v!=W#AELCMsbB zG-jOB?Y@oX=l6T0S!>>7cw(QwzHz3vs85zKu-_#j2s-#$K0tmn1zeZ$X@uS6Ph&WL zz0L!5DB+tb0~WS3l^3TdvkfcJSqd zcQ%@zrfH*wqUXx|O|$6mH&XnjqN-eB}PFbVLN^n$!zJ05E z9C|GKgWj90-`Mr_c9m0YV7;&I6z9~^l^rs47mW%T$s8|W=^&kOo!sRwk~dgxvMBhS zue_VXo4_Ag?z|RZMrq`K<(#DfcuiBoF+6@Nc1udqAfW1{ zrsJ|gu^8?Hq^lI7nX{1R-3>srSuj`r$<ec?5GPM%0Ps^=0F4c2+E2 z*V0m7sY^I>@RSIDbH#U9Wg96ss4eL4jaZN-57LXX;vBC3_kQtX092M@=|}VBsDKsp zM5_N_*>=jl{Yp?OZHwLk=GHm=IiQzh2fIMlPk8csi3KNBfher@L*#uKzRMO%e+`$v zbpGVAM*`vlSmO&E$PG`%hx3Ckhc#ABWfDfj5v=IluLX}chcr0g$^0SMf=tlC#J`6h zOqBDg&ap7F?Zba`s4-(sGrVMuVsFky-@d3--~R%5;i12*Ve7B%n{y#PYR>%gBtyUf zxezSQ{7@1Kp=XbMxA*WD)zjoh%vq6~!8|9~HvP|;MGTK@x{BV93P5SCH;?#Bf8XOf zyw4fi{l!475ME^wS97r9>DyevSklw;1|j<@RKxnY$k$qCnN*clMd5=)LbhfUy)OTL zL>14=mzfWpqQ^Xoy`uI0!hHQ}41~%#*z`>G41%vhKAkO3a;v~}|JlMk0_yqiCjhZ# zRLe6pmh;ljxWmA`2Y@)VhcE_Q^=VZ7N9z35)aKQ|VH>rE)U%LYd_`X<3h(QvZ4CCDETOnb9X76o9ioYfXIW*KtGz5YA zI2JJL1Ox<#KqPeobV!CmDys}pIS-F94-eLWh$El#tocNlReYpfZ`Lz9jH0V0k}i~0K4WijSI-xIRgh_$C+nIv zR>OU|y~f8P5=Hw2XN5JCjvVPFyfYu=j8(+bw~AkjBO7GF&MEVqtzj+HW{)>Yc z*RKr~9e!;QLvp(G0?93LhXNxvF_e;5uV4^C4dtg~X%QRLGjBgnq4Gnlr^t_qs0C>a z`^?*dU>csJ%|c*6Az!_MBms9~1K4qBOjpd#S*JC){NEv!0FBvD+|>&9Bob#qug5$S zsnL*EN6N?e>)w^=gXkQ^q#~2tt1Wg^uOh1>K<9JA=1)dg{dNevA|M=glTR&0LGrM+ z(Cm?h|7H9iTx>RV)fy9((7n^S5{00B(Iha@)8LeH^Z_L=dS_aD(14X5F<)K#{!1+pxdQ8QcB=^0c_kdRMc<#-#_RZue{z zwE6$!87VODOBxkSisTId>*X`L6^_E6F;mXy+XGFPsoH)TW&Zty{5!J<5tiqF6lWW! zDCcco2Q4?ZR$~DekVToP#^4;=H?V9E8-bwp%s95|;ma-UltT59n(y_oFt0=ar6Ja? z{^vO~S$4&_s~NLjDSu0Nii>sqRHnlNDNSpF4t`RGm9$+H>qM9WTGN+_Yywh2UPfYL!83marvX08eK&6peZ6aO(j774}E^Su=w<1uCpa+p>^ z+kAiYW8D2BoAll!!Sbt&JcBN|y1*>BfH`Gj_=elcn zU+NnQuGyo1e2IT?bH1{9o>lZvsi7*C%G#=hs|yq^i;M{HewJ&hwVAEuPQeBHLwSfz z_BSPIRJ!agT`?zvVc5G<))(S*5G-)Rs79(SXd$A@f2jt)#}-x!PU*%D9DIdqflq{| zg(T{>iuOJk=rfDBPJ`rLR=@&4sA(GYQ=3pn3%|vb~7`G9@r0t~I z9GQlZh+FgS6t2V@Bml-bmcJ}!oGptT0=dcy@67cPr0(HNqeHy&y+{oc>kP!cAg3fc zG^xPu;x%bjBj^2`FtA^-ZGJ_SODBhW1roY=0!8H7P3C{?DzPg?cy)<4J>xOZ1jC$p ze>#u!1pK3ZIIbv!{KBKNpNL3bhdKFoW&0p6aIfRspD^(Ffi>w*DdJ2?gm30+3|;(T zY=gfkB>=iJWm>Z;ycv>jFi9E!;!SpLHKWaXKs}LKz=AM~m51A1z{{Z7)H^62^sm^S zACA575rw5QD`8}h<2=V6;cUdu6zi)US&vU$c684#L?TcRL$Pp0c9YS8gh%O~PTS#% zK58^3`OeXuR04kWXK3oJ4`P|H9PHd&37e`n0|N=Qac7G?yps%932**?7eIXP@*8Jq zoI7Q=hHDTR|9Q6ONO+z1@9;hV$YtA~fGOkw({NNqund9w>g_0!T}$!#0+md*X12aW zqKfLX8%1i6`bdI9ps;#1RS4e<)}S5QnNrvp@t(1zNooHuh!Fl#^RPy8F*&+kzN=plG&gp8kwjfOw>*Zb0@ z^dAvlhH=()qS2wR-bz_|?i76hs$}dw?q;#%g!Sx0$va+Rr@8Hmi<3(`tbK_cNJv8x z(qsOI-w&zeoMjh@s2`!b$@}EaFV2QRTX|d92X?%jckmDT2R9!0idboYXgMs$R~IL0mw{#n z;W^O|P`Z4g#u`M;;R+x5XFuRHV(r*;e06yMKncKjWTM12+x5c<#9GRP3ngX^Y>n&` zxj+Aj}Rcn@V6zNh`s0KgX})^VuLL&F~78n{dUPN z?83|or*JcQta#Lgvr^x8IpE3<2gIBHnlt~4(odlyhLV?sbVu86-ombKtX!~2HOy{L z%E$E2$=fIBXS^)b_(AXK5#3T^IN}tbbFoLL-9$DI5&}kj+kdyz4S-v-g-IWSD|ZU9 zc~1+@h(_>aoxqmi<4w?tX*<%po{dI_rp~90jz(6SI&R8-HDci5ByC(1rWm~yZhDmO zW?(XxX1TuvrQequP(;KNP1w)onBJDe$(VCWLRlQmR_JL^P45UHSbV2TKp#FGfqwTz7O>O}k8gY`Q0hn$x zZ3*f6Jv{b%Hw=h^?XR74GVDfer_W`24iyLL*n%6ZDv|_%7ji}rcNp(*4u7E!lsOc1 z-YVS}GQz1}C*ceu+GyCq_Ve1}LNp-OdV;t@#41J*JT#54jgHH)&@l8OK59l{z+>zg zhHJ#ua4bf%M(K8$a@yQhS_tfYI6YqEz#*^?=0Ag^u+ZHxU7bVfW#qc9CYa%RyU zy{w#{177F*|6Z1g7OF+&NkVP@y%w=%(|c#-7F32U-ix@SUELobmY$IxPL2f9&3%06 zaH~tx{_#*mi+T8VR`$-5*5xzYifI2JT{*4XuC4&ZX)G1q^K|h}80$v`Fb46=mrW35 ziC{>e=qf78maxGqKBZLOgKi;%q!^rX`lBbp)#WCu#J423Oo^L9^hwE}rc_&HP8v3l zV0l5%wKAn&g*ceJr(qXRj|inUnb|etT)w*8?9ou#>FRYayK;KVYrhc@8dDE^yKd~@ zezH;%Uxp_AGzy9>DyUBFdjc>)gWL$C7_*Bb#?+rKZ2Dhp3386!9GJ}lS%PYIc_@3ExvTV}T(Z&t|HW` z)~%2Gsb=oF7VCxQF|nA*#_yi8B%}x9}`% zMo1+i{s&X??%{$F_hqYM2$U}Pvye%^K!ol_m zl6^5T9Wf-UYY9Q$Ag+tS#C?UJq=9j6bN9N@`jn6~JC?p5Dk=nKx^=^;S9x^jQqQPY z&A!%T)4-5s&)#p1YeJD)#BvXyO|Dm0-vhWOr>104amc_@0YMIoaS3#SpA*T5qy@e> zRVICxsLbjY>1;{!JHW!X?_E+pj7>{2Y6+pG!wq#BI0xrK5Ay14rix^T0~N=wjC=TJ z;PM^18B{gZCHK5nNi6W=nm(7JlZ9%#Yg9HSSY7CJY~QnS%{X`O67Sutu`~TYrY;to z1;U77k{UPd?2k|a8Y6QXWn%h$Ld3vL=@n)G<%95%8_ARcl2$#*G!g2p=U)AeB5g4n zA7RQsE8#QMK@8-Q)DM6R-s8zfgXNLWOf<^FMxLCZuhU|aSmC@UR}I=|yF zM+sm^HI!<^^#vluoP#TD_UzsiO`glQh zB{TaDYOng!v)|`UfWFJP2L4eu)IPHAJ?w6hr)&Vh&suk&*pIcIc6 ztCI~9;@7Homt{9=?$i-4jSJ}dastUocnu&LPJHjq=YmSh+LG=VmVEzvF5*!>Htr&x zw++K}8Pl6bHIwVB{Kk!{Tu``>*eTcX%Aa_(90Bz)RME68_1*~Cy7m(l03lXQw zOjM)&_8TPgk4p4383yOUD9}zn2$XLK0<2wm z_2L+mX};KMYF5h=F+sA{+(U5`=o1m`FOyy7YK$~gGC3;o9O=%wP`yIt=k<5_PK*Ue z0FM-C#FCkFSAgT4z8E;?eIJ{CX@K{y6Aj$~qX$p2vT6Ust!Y91{2>et-*?PmV_x?4GaBtl&MKXrW;1|RuJ!Nkb~?1{(nSC5QoHolY5eT?6O z4O7T`v|bc9#ek{jN*X#Y{9>Hv2S6?uyKFI^MI+RfJKPEr=-<)B{f=ulgSz_t{0{7U z5TJfuY4u1Ga-72vV=sba@oJd1p|~I<%^4(1Y)>-@*W`0KwX?WPO0G*R2BHJ%pNi!# z-gD^-U&sWOrWdrbigsPL2J{fPFH(weeJa72Nw~5WherKY7ng${ZeK$rODYtL(J(VI zTQL=o{B{QXAfmev2>16#sg~sJKC(27Jrut=P_a7moeUnKg2I>4>7d1+H+iLd3{sDoD_Vl_P>p|UTPsVp)QU-7fY4c%r~8xUSToFXXP09pRt-V+G`fp zINqDwCiH!*g0}}M=kUki>gCD;#TMDv+MBHos#&|S9hP|S@pyr9RVUiJq#!ZWAOj@x zwFGviG%(9*v{4zK*Kg4r7&9MD40^axo*h;P^MudPn`*W(FilE#FdmtQ|Ktk8u@4WT z!!gBkIgCuO-+&3E(_-x}D6BUHpzS)?W^SHLc1BT|srsfXR zjlgtWnwvl+OHVu!CV2y@G2}vF#0Q1}#&YF8!ZBRv3LH-+W$AR&w7?$c!ppQWhojEblEzEiQkuh&U6J%LVxrD2neqmT8Pb$NKz&mM|O&X35A)K^J*mTln%w9C(--dczRh(jdDJ=(mtWhp`Tf9?Nx7 ze+3UdsEhyu2BlWmVlN2+?j4fRD2Sl?V6E)qgTiIk{93lnLNaj6ORwQ3aWz!RFJQX; zrJSv;j%LHGu*()~`jQ4_qT!0wD-+}QE$x)Q^CT%8NEUlUUm0uj9mWI~N8BMC z!jdl-?0M1Fo>}(k>$mJ#sez;KRq9lj_GASNDPfZPn(8>?gOV%O*P%#3wv_e)X&e}7 zoZC|TmRHJFX!;@Q#dB2aS17U0OD6XjDW&W8n7#oI`L)ui2A*pxIN_K-5eA|!))^$C7E);ca>13{uNoO=TuTlEmlEsxFCo)0 ze^Zz`I1I3NunElMbcn7;zC39F^oA(NwsnYRGxFkE2~YnYrp`N_t@n-lM(tHwYlLc1 zd$eec7)7gUQ>(V(tM;C;H?e7LYSrF*k60D7wFzR+*g+&uet$f#=lT2mm2>X<fzC&<4&g2VQM&Sf(5Yb0`}7-_KVW?POp?8}>_p?E;=tbuoflMT_eW%(<>3?f zgz*Z$Om?qzFLtMNO7^{BfpPRFsU_CmwHp}6SH5VvHDk(U=uXEDlpmCCgbE+!6$9UC zdmX3D)i!4wy#BSHdrv`&d1GG1W7yF4{P~cgyrJsn8=bS2Zhvo{DDIE>fV(}aT6=`% z{#ti-;&MWSFsQ#p)P!C=^P~~##6B2B@!4j6y+5ojVUXwboQpJ(IRbeW1vs0TI&;o2 zH3PlawvgKdr(hmVUG-Li5b{!O58#a2-aFN|o(JVXwjksg5tFZ%d`O-So9~vzD&A3J zVy;WM8SXx;Hs22yfuDE(7&^?LpC5oCNyFpxMCn1xGdR#6GIt<@9Iw47o`@2w_R+QP z6@3h*eFR4D^y7!bVqbM}qKb_|&OuMeh2jeBj`0=sU-)9J;(cvm8f!zpdMp&9DrmA| z*Efe%-vAYG>_1Xx9aQHl4*#_ z2|w2Vu)GCgJErt=yI@z|i%Q*Jcp3%z1-Ch>8n*Jc>UXwT#ERSyD1FKiKK;Wb!*Y{7 zXIeY$7FkuwmPMve4>ILePKovDd%HcO_OIz8gacW#-NxP}^R+W0P<*0&6CJ|IAK!_C ze(;`lV``rMXBu$(;`NT_@hm;*-9U#7jQHqNFFBUSLW>W!ov*gSRJVE4|8ZK0?%dGF zN1*a(19vzIR(>Q}8AnzBDbL}el#ugPix8%GT5aY~tgMo8{I9sI%Gj#q#U`>0no~F) z_%s2V(XZcXcz{0`@;$0~wq}(`;rkf_-Sbzae@DOwZDI9X_o+92!MuN3EdBYXi15cU z-2M`r!zo#kI$Q^pUz@i?f`YwRwI z0$AVD85b71K1H0%_lTy2&xDMh-WkxV!e1vt&X>XRhxPx=)(} z&9K97^`sB@Pb8ekt(8QW2oC$0jC6ft&-Yfn*JfIlf)F82^0Q>BML`dDAGs0y_w^Kn zTq>AvIp-M6;ytCDfWzT~i5!7eBz~o+pk(rXxd17i(~4J{9D(snha)|`(Ia{ywDF;a zi~ex6W3`d%T>oQ7JzbaeukDNz1jIKOY3PSF37B`U{}a5-yd}>ZkYHHW5n_CK& zBC2sin;Sa3#&lrRjqwz*cXV;TePecFqt^daI$KepDt7^`L>Dl$GXus-;ZPLGGC=%>$0Z=3CiS<#bs9f zDQm|H+;p%$)Ix`bvk}dJ*Dudhue zT*V*XVxlhyQCq%~80ocw?XisS-`HZaXc@&!`k=(J zv|&CGu4G(@0X9r>T`AQa&!DXkun)S~7 z4M5=?Kj^7MvMzN7-xx-d9KQeA&mIL~^AGpG1Kz-Ss_#v1YI0c_gNfu{Dl8#oima%2 zGStzVU9N_Fs_4&Ls73Iz7T@$^u$SUJk*)Ma&IR5^=aX^bW7g+n?QUF^(QO#;;0R9K z(58$yw-zvtkKu%l`kP@FzdI^Pb5MZB555FpkC33w=xh;vKWK45Tgc~$aW-JJbd|N#kJy_y zYouXDzHv^#`Af;X8Ws_kkA0wtKAMonqj6_(Q*GD!dN0#P!?Q4kd;+*Zd{d0Eywk&% znFG)k8G_g(Cg8sF(8rl-99FH$vk*`D7{gVAAd*`$$@E;j&08~4phG3{*&QecB}9lZ z6l%H|kIFaUTNg&cmvy^r{kG(kU2s*&RemnTPu>XozNHn=c8wt66$Ce&L0p)6Cr|T9 zKqHA+w()8xM#dCW3A#)+<%Y^D@fAAh<>m(tt5~1(5!mbf`?BB9TM)EEVXQ9#Y_N`l z^Zp8hJcYfZu)*M*{uu|+NN$<&(J1f5M}q@H&h8uJk~o62sGL``ctwVdr48hZgq?N! zKDb|G3}X3*HjiE%z1^;{Ke;|`QzSZ4{{lI#8&r1!9<6_UVTW7K=9M)MaawaMa6vLN zv~zbqlx&O#QR3mAZR6P%-y zT|Xae%=0#O?rxSM zx5WkcZfz8>?XD{L_A6no7hX?Fv0>xAj(fgeH1k0cU!S-*lMoDIGc5Pfh$x_4eY0b> z_(@V2WyUtqzS$mw9Y=4_!&P(NTGRf{621CF2$jnuIivbwWc9LB6oHt5k#!=ivhE66 zv2^_B-hYi>j3w~mW)tdNcp1k4-4{Dk!!=*g4RSb_P%~kTIbP4hnnSO>HW@m)-*tN^ z2CtR$&hG2QWmcEzkPfS#L19;Q3#O3fZ;RS-H!4@s3R3OuIzRMr?vJ{`{^U|UyOU@m zDW#a~z)1Yby_uEa9*(6=^P?T9KAFf>mu06;*1e^r>SlG;(1DPrp4p^h)9t&tM28ue z_9@}3-Kij6mhyrHh76h=UX#j<>Xp6=qJw7!<8+Pgu3fTe&SZ6+{+A0P+K8N>yC+VS z)h#J_E}xI+F#Cy8^0Kx#Z=FGcY|yZjxMR@mHfV(M)e?^Kj}(Tz!)l#*tUWWR#u)~G0ovlW z$-7OI-X!L^SZagHVpe3Z9>9)I^@A8}&&I>JfA$*+$!J2vzmy+>&=7Yl8g}EI7VO@l znSV4cdg8V_oXh;mDL(fqI|d`IUL=bhR`nsnXKWbac%gvf_9|@y7!5!=HJCg=7_R8%m7|lQky(EA{3H|FPjE_0KH{L zVy~Wid;{?Yoea#-tNEXgm8G2)#R;adxCE4w*)ZI{@4oX~rAkWo6mfD{X;S84!PuX< zJS0f0M^{L!lxfzS+a-!S8e?ZXy}UGJufo#+M{%{?=(qe8rBaD1y1K7J1|rVQ$QXoF z;6BkBB7|D|?{6InzGTzaTpKl8%ghqbDQPI&z9*C96}1;pegR+=VCU*V|60alFqS^d zQOzm+Br)jYgM)1DO&nSNdgz$8`$t1I6))B!$@alzl5RO4(jk)A35v1(a+!%$77t%7 zfXo`GYbL9Sdj5u;gFUUFdz20Zt2cXsw-{I))GEM;$gQ#V&|kqUTGCgKYkU1wldJLQ zWlgSBXIWmr(?A!2G&2p=w?@^jDC1T(0qY)DxcSTXhlxSSjMwD zb#D;;*}S~Cdzsq&q&iSt>iCs`*Huypl|cQdj**^YSp9xxFEV|kHNo6|5bY~ce*3`^Vci8sN3K2+uQR| zP6FrK_?Dv&RB?mI)HfO+;@Xpoe7%q$hq$WxfIsY2U*W|rC=NNj3k}|JJHxj=)8>YC zUdIiU{hu1A0MkZcz6na5D{(7wC>IP((uvHsJ6&}h~7pPJU@sWA3EWh2!kpc~YY#%+tkSB<}8Dtp6l z{;6`@Qv|b{OY=NMZzGcA;yX%f#^uUo1?tMQBY-`8F>-f*c@;$k;L5AG8hvsHM|WZ@%AE8)z&HT1SP!{tcep2k#%yAZCox1k-FT zt#p`z^7%QF zq2v+AW~88k{P0%idDBbA&03c)(@=`uW(@YjS+ivfR)k=FC{!+|z4KE8#pi31Gs4Yi z(2Nt9BBT&12xWYWXrx%2xK&ZvLHv%=da6zL5kW4Fq__VK;SQT@N`V3m?#mk(thZ?& z)|LLlJjYaE9GEUR@y%E!9&qHTa;|9a($YqH?cV>BQ=MaBvnhNmOK$*exjQp z1Ryey+#+JuvP3D)ZDS{K4YcZCTiL2E!!NzGDS8G(FThWZz1K#t1XZyq*=oo=V5{x1 zEIK{2F1neL;dlxDK5xbd`Cpil@Sw(zT!vv&A2b1f@La~MMIg(JnC;yB0>MK9gL6Ai zrwgu~a&CqB=1^c`{x`xknKHOh4_WyVMPYI--IH_ywm;!UL4;zO_tX?U3|(O*A9LmF;c|uZ5Sq5N#yk8ryk+qg}_N< zVTFvnzO$s~5)Zj!N3|1*h=WL8;;Cx9M176c0?-xQm(*7F7FV(QF$ z_!WI7>BMM(VEiQ((rCtf0M+)B#JWL+dXnxLjgl8&clgX#wSoXNPU%Gu6L0SIT5Hr& z$hi)g<4|#R`E%(A<8ofHXt7Mil!fL{7XlorZGVe6NBtbPG>dSl`Gt84{|(jaryDhl z@EI1J`_(OeLiD{PV!w8gF!~P@!IRdEy5g<2KH2@>tu`bhR_9+gCAo||n(g!xd$g8+ z6P>o<$UY428>Ri6wI<|;G1#FD%i+ZBJnDuIhGqHnJ4GQ6QHP{W!FKIRd@K_;U3W%0 zJn#W6-K&D~tt8M+0bR=7wTo~e?O;FX z=Rc(<(&`CtYF2^vw;n6}e%2jANmm@z1K zrBU|m!be3m-*E`vae%HvAbDr zlV5^eBgZRtXH1WTQp{$;$vr~k$evIcW_hQlejhQA-PkQ~=#R&!anfNz#d>9`)$)y= z;EL|N;o}y9-GcSH{7@n5KZ-zx8)YdSQgorUUFW2U!wy^yv1R#wII&GrmiX^R3sQjzPj=z%$L#eOzK8Q`nib50a*Mf52fFqOwb1RYa%y;|eR?D%+HY2W1i?n-E)H<%qCGgwgu z-@Y)mc}x{CbQf&JievSUGkyk~`I*JQq#*xHUr>g3viGrrd?<0WiEzbMZ90uA{) zj(oy=(fj2-kq7d4a)_|{@Fe}DFzsR*Pmh|rY{j~;8 zAMS>VpZ%AdP7UO9N_g&yp!)aOlMA*!>8ugzAItAN3c&e;{HQ+qq*ft1;5k)-#iqq3 z$Xq1_=vmuU_8n@?4W{@qRi@7%j1YkuzWHwd0X_QfFNU4W9!0Whk?<7spFQ!kq(xXn zP}fYQ`KxaW1{+c2Ud-+@k2kqLL_^Pa<#MVBR9GT^LwLJm?UQr`zo z(5OS)>=lFC)PbizFQ$arT6Ep?AX)oF`MhwCM5<5@$38((L^4=e}0vLvR z!bP`f1k%o<=>@HW(8HNaFJ(RUUMxMlw%W-&k?4J+PPQd2(Zc36d|^kuCeyF9U8_lW z>f}QYu{iq{z{|_k8#Vm(D7I^}7Bq7v5CU6v>K5?I{`9Z@leFdD>#U)RngnS50~^vb3*$$qP!+V5--wQ#HG-}|LE1ujGDCeQKeAtfF&xVgL!=? z+sTb$&e#lM@*1y}Dr6WR9KHGQ_|&35WsYX+HbG)7T4_V5e-P|>MCRRxp_-%2G;#Z` zcCl>f{pOUC>6K3bLm75{TOj?Gje*&>>yyk;B3#ggH<9oXP(aX9vB2?R|JGtHfK}+v ze{mTdjN3L%61UeU#;0Ud^5w?$<0_%woNLZ zto`$P8?#X%f;_;8t4qSS!R|-Gu;SXy`HQ8E*GHPW3LC}->{}<@jZ^5bPBKV%!V@uN z>H8zNhG!EUm=>P<2w>@&)%nI3GxV56e&pkq_KP1Xv=q!gV@t0#PlL#B8*gn4yl8v(5l9^<5(I8|H8q4{xaE7^>cXw(1{i@4iF|@({i73@L z7u(H{l~>+~>w8`E0MwYu!nP@u>aR~TW~PvOt7ccsgzSl}6o!0`@^@M$YM0Vw9M-7$ zXc}5P;?ql1P1@OjBsrbzQxk?s1f2A@OFo zTdPh=ZuR)NHi(su_Mnc)44hC#40IanlH5k_GkC!eZC5R8#nDA(=dKF)+?qe*C7!n; z4T0wZ zq>#AQs|{A5B7Jle@b!Um068E#$(eb*XbWh9dxaK@)?<7h>PJ0TNBUy4X@2+lyt}mm z{iF$>18m(KJ!z~fM^`6ySYja052qmixy~o)LMRtj56V@i8FMqHNz_o7hFw#2}Idc(R z{Hjzd8OhrNI5eNM09I4o=7kRb5(9xskx;RnLmVM`VO{-mJD#}^?{iJ9M!|!rGj2(? zCn|i>iQcqz$a~zh&@SVsSXo{U-^li*0+FTlX+@JV_1v`A0U&R=&c@TTlaxQ|Qi{@8 zgh#^#&|jY-?`g6Q^4zR68kNFf_+~sf3!fAN)movI!A~jT=mLa=5eOt?QPT+F({E@H za8Ick-w1tcEP8Xa;98gSRuPp;q4XS*MOE0fd6VuPU)TF72l+IS|9gT6j*5l?$%O3m zBYnHWVkb^lN~!Hg@8keHg6H%ht1cSEr}rH@OAh9N4X^t@FWnMJeL|w^aPcsW2KuIS z20y3=C9lQNdVYo`42~BS!kEzu|GL*W7+nM9qqSS12&7qi|c<qambtNBisgo zNR8dTA_wWj$-}Mpp!~J&*RXX7nF21G@}mD<0@*&Gt9q?X>XpT=-;0kvZ)w4BVgl}U zDs0r%=*_~OzB5c44QMYejCnc2p9s;n873xzzc~}Xbzm}-auI_T*FD|w{*jps3~eWZ zp&kl-y52BaR7dACFp1gD3}m`TD|g5}+&R?zL1n{)-{AhSot>C$l5TAxos=m{{|P%0 zsZaxjMYb3YlDcT)$dtZq$OCskTypj}!7`f-lGvHw-w0+)f5OU~;Vn?-_3=BUK1T*I z5s}>R>Art6lpM?NFfms@0L?|I zy+f7p3Uem_?NOiLLQ%wC#P~t;c>kGLs($!Jhmq*3~F>%!F162#t+{Mw< z3yUMe8*7m35z2#XzI216UhRV-PSW$tJ|Mo-y~beC4hiJup8_4UVs|*qB)(+g0&YvS zFsx)HD0;R>%S#Pt!{D;S%qpvdrv3Vq2P1_$`51kzia-^53?Rz6PFl8Id2AL0CL*xI z<2xj~M$_5|r?_7%jDtQ}Z@~`VLjfaA*cb=^#DhP()1}?Tu^Jp5jaj*!VD0XzWCz=- zyAxcg&@sMlCmwlbFz;K=$yogHAK2RwML_dS>+LEc3LU$OC{`Fo+SiK$ z$T|&Z|Jv*gBfqD{tt=Umv!5u`Oby@W2YzK?INTXcgPj}j3P{rpeVf3*+ zw&+XGBP&STDDtSM)zDODJ?cq3gP_W2;%$+&33s$R{`}`>{6XmT!dZ67(%-|D?DP`N z4)Zgi2hS1dDu81!!T3V5ciJbxyV+!wpCQMJZ=<6bhjc6oBh<{K{%L{RUR^+D`15a( zT#ohd6VGu%#sWhiFWM(SUp}bG;R$k>;L(^B3CaZ7G%--`o)<{|o=5QOScp7X6P;8G z{RXmlW^ep~19+%UgevYz4sy@<#d<7B6Ms`k*O>$yYIHR8uIybktEXJz8~ts<(mB%e z+b;cKSh%Qxm0I^nh zz(Vg_#V8ofJ#`AD_p@tIo;QiY_JVs4f(W`j61ueX^kRSPuqs3cx zSDHtoROf*7V&Na(8mjlDb8&$JpRp&sXo@^WDH#9o46i#bqVc>bOLLo(=UQtm`TXtO zGn*L5q855vI>-p*s>YGnUKd<=E>sG7Mb|NiZ7+(1N$mTKXu6Iw*mN-2hMSs>zkh<9 zl5*aPj;@IgnkkQZ5yfp2^&o>!MdH)wxOljT{?{IAnfVwvAuwvjTEsj1zg+@_;~KYo ztLV)0iPEjd?(v*9rj1b43q+&HBORP9MiV6u$~s~*`5?l*2Et}t88RG-cqng?Sg=;H z^1gcIle%xbQ7^c_g9mf}U?kc(fQYj)xi(0M)t7QLCeS%kpPO6(c>}|_abUFcP0hCG=i$fyqq~;L(@V~NGOvg$GO826+6lMt-_B1hg22JTjU9NLt4 zmGtq6)%^uTKdV>-A6&4hRY6xDv`{DdjY;S^L;{~$mqlX4wR4gIabZ`b936KCej=r> zA|6Rf2e-}pEwMa_fRDvZ`OdSFt+lF9t?okn8e-VwzjRpYyeq)q^RHl7RQRXD2c#s8 z2x&1`D@(nFt=YEK)3fUVEprck@a09D&{{I}ALu*NUu&(WB#+WN*?j$Zg&+aNy>HRb zckR;uz5bgWf$$l*)^6ECepgo-9afAi&<)!edF(j;m}^)4oy*<9~i`w1goF8nk%87VFP;B5DY z#x+G^C7?-QXOchpQbI@wq8u;r9Mt85M#9~1pT^#=sn4wnM&^(6PPK*|Dtr$WRJnrr&!-`W32K``3s>!e|7 zPUSYM`g(4!U&bu;;pk8l`E;^~TgES2T=Vy>*Q0$Gd&AfCbQ3sV^b1N33~}tt1N5nL z@%aZnC*|XIgu{^Q?R;$X!sFaF&G-cvG*YbKG*C2-<1^96F>^;1JiR( zg3!&yu+Q`}Z~C4E5Zpc~;Nt+9$k#W8vc!(FN#`k0)@tf!v=M8V@^iI>zv!rX#*j=P zMg@0o?K+bfnh{iC@sd0-_c4+30uMYA;ZZP%V*zrN8+dGPl6&*H&qS*m$3EU!Ws#bNEJB(lmpxe9B{U<$BiEF8jDNV-*t6P1GN~|TYWb%*>t*KwUB>@jdw99 z?iFV+ZaE3xVUN=ewrma4??+Mk7! zWEe3!YN=8`i!M3?AE(a4>^Y<}m<8{M&?y+BSA^Wejg)BuM;PqEo8UWIaUKOGi~t?0 zv)l)lJt0sbt5|H$_&%ogAta9a|$ z9Jgn?Sk*XY-Q%MXAy<$s>`_(wa?|aW3<>kUkuhvXOl30xFY0>=tKYbNpx)fK^>dLf z`o(1X9^NT#Pb6c3Dn@O`K2|Werv+B0St{Y^>{~xq&>%s58^k`RP(HSRXN0=_BMbyt zPjfbQb6=OxafI_aVUkiK+)&BP`hQ& zw~&JOFtXu~r3(d*eL$i;*vkAG z02~;sW}`$&D@XF1O?F=LyZXYpTYe#(?R);6ksg~F4E|D#Q|#U7bAP$ldp^$R%CGOu zTB1xas-uB;U-d;#W0PzCq2ZSR3vE=iUeRoLMjum#)t#<)Vj$A$rRiz;e@5R&_r-Wt ziQ45R?e2|5kWc_IU-FHqTgON11G=+tm56%EA}U$EblkY$s42Y4|L&P7^njUZK@MUW zwkS&&pW+}NTV=;);q-0&v`SsNSdTKQIl^4IK)MT=IhH?O|82836KJ$PjKq#$7vZ6F z|MNI#isL8n9oiu=7!`=fhgc3HqrS+j%mXW5@fH7fI@cceAMKtL(T%*TYwjgLY79IkIPKAd>z01OGUwU1APR)@cORImmIT;p48)|rWSv7-QW#IH)%X{7Wx3` zoSBqwd3mq!7ORzngwKfji&hopKe>-j!)Y35IX_tg=F1rQ=y|{^W zlV85430b+^u&+Kca9-->C)OP4sth+fQ+@#^kJ}ew!RQC0Dxbc=IIsYHQ^|rQhgI&- zex>Vrp)ybMRegz_H|@@U%018LcpV55Ro1h#O6h%$)gS)F%p!H{J<4UDCG>8PHSV2i zu`uP9IgW$AJnl6#P8xK6y}KmHb^M0h+?~`9EJt$^^)Zu?XdN!cA?N>>q`KYMx!@ z2j^xPu~ZWFa$IHi;l4A*V^BbDjDgkD|t<1p4< z|Ksk~R_U?qu}sarYyKo;e*H6#FK&X)n+j;#qjB76;b@9bYEpU0U?h#GG(ldKTwfIV zzjZ1*(g}mxGYlKVYpLE~D%a~jGw*G)NG1}LC55D4z%-=B)%eGnpPj*EQ=JEA7eue4N}I{=+E_oM%)T}=!X%$yTv=zdY>89 zuYN@j*^G9L{DgXV()yIr{F73H7Puqlo1SP_+&teXv%uyPyr1;K^c++PEESz&0bVaT zqaz?K<07;sy^YYBTr9I99;skXo_N}|N~25SV@NTa)<3XrEpc>U;l#iX%$np>lIo}V z$trleGn3e8?m65pcADOB#8H+mVlD@I$!@UniofBMjEvOu@fr){i~5j)%o)+VP3u$x z4GA(Km4}L;@mqTDBUG7oSS;0?A*B)pgy4KI$}EBX(BAQD4Cz7iZ)|*-&1K^So69qrdgQ+t21VT-P2UZWk?|Kjmk5UtEBo%xLcDK%Y|TL|yZ#Q6zl zNam5m%ZIaK+nF223CEdHYF|e@&7@e$9b5#Bl@+q;zJD-=0%stgm&V{djv8*{k;?N4 z_Q+7(@W2oH0SxaH3XN4!EBOx`Q^UK7dD_S_knrkRgM&WjG|>0lnw3^hGo)k)Fsftt z@dMENgH>Y#FVlirO>!v~zCZ9OuCj=|JzX@8&$r#y$P&h#VPh{cs~ZMKqX{H`MBV!K zyY46q|LO+<7$mli%Op8vz3ERJbLOEReef|CO`}98z{a`DGBoqwTOtE;RLK&M&z@v2 zdP&&#|7HPj&t_Hs3Wq=&&w7!YVx4gzsW7IxAJ_SBJ%1!VkRZAyHX5@u3NR02r?$UY z7@6VRw2Om2RkUEs^^()?m4v`EeF#STJTv_xRCNF&2hwlN(gKE}OV}{aP+#pEEFrz@ z8k-Q(g8cEkf>Ua&wwHbeqrY=)mT_rp`}=1=%pG7mN#C3c+!KQMRf7>CyTx9CSlYOi zF2zNMvN7P-;M~?e2fdcY7|uNO(mZ_Z!@%z@$LKgfw--v-nt~!e8-a|eShu*98?Zcv zwq7eN{HPB+;~ply=-dAm*hMe_&a!6T+RlE-Z!m2n+~h_-@!tJ)bUq|p=Hc%{+irr? zB;U#t+nkX5dBUh4i&BB01S&dQ2Q{*LwC(RW-&ad!K|hz4g_b#kt76Ib^@C-)7uMNM zllaWuekR`%8r9c-+=vJKbL%}X1M=S97GUC&ARt>YS?`5kOOeuX=spVPtY~|z4(fw; z{+-(QrsP+-Q>C8m2_#y5QaUBij{eNb*EA}7sV(*Oljpl*6D#I*~!1I9;Z!9S~_ z&6$UgaogZysMJ!!wTk|lv8dZtnEw0cAQW|b8w%=R-Xps9j+7?uOia6p;8w^9^J4@& zUJ>+MF@b5U99}|LkG7>BS~fw^GhUMGp&^p`pdQDNXA<&99@z(mx&Q zhZKnn2->5~6p=8#ps{PFly4cQNwXABqlKrh*QI@Ks%MDN2+_Q_uz zjVpQ+Q50~!JYCAEmx3`B7K&QC1;mgOn8rAPF}dDruCR5V3=T#LSD_w+5QyMZQ_k8ryxi?d7?P zsQrP?oaf%)7NWrnM@+N2ogsRJO9>V~oqv`t-`eP_^g)52JoT}2t#rk3hP;-<7^Wj( z_ieIHL($YVTJXraaG$2#ZVv323|=yEnU>F=lA)cS+KKD)lsZ6Axg8PqPf0^ z^{21bLM=YTD~_K3$e;M7FFhX4HLoqEX=C^VLG`vbpIH*|?Vl~8K;Frv>hTAdnRuy5 zKhmlwsU*pFZh(uLJmUIuoQ^3%$bF5^Ff@Ntg?bzUBB6WMc5wJ`E%I7!^JXTY4)H*E z61xW;K6UG5SC<~rXl1T?#1?p@C6LG$O0^CTWaZ8cAiX0Le0PKV?tvw&HrbM*_97IY zuz`|5%2#hoGdr;f;V$jIL{y3kqAUJ0^f$KB&#>E>?uzl*Sq+fkG^Ur8VU?bBqfC#4 z@dn_Fr2bniZ-e3wQ}Vs5PK<8=MHc;s*5MeC5{~Y5XXHmVRy{*?o9D`xj0FSpK0g7T zg<|lh4uwk8$qWn)zl;~)*gu+xDU2?A_N>zU#hve&%qhvs5zN6mr{~3FzGv5{-s>OC zqAwB;$a#348(R3yckI3gPSP;^=}(7TNb3I}GfwR8=9!hT4K_tYn%6hD_Xud*=t-#dL0Meq3Qk z*=_aJtFQW%D@J%JJNEeM@LNM6FWN6yz0WbhAH85fH;y;XCO?9ve0Te;{aTj0_P_F` zG8P8ithUzJ|M|N42a+Ga|1ln^-FaJ=mKFNto{Ukqt-57Rn=t}}*dRKQGRf=S&LD zy6ku8xS|3G3~yYTcgqh%e_ytY6KLWI!qPkrQ432B|6Fy-)CM2u9NI?`B&JJ1QWy{0 z<8PaSCAB?2aC^QTi1)e5U<0-@D;|^>DQ7JGnO?2=-aF0{@8`@<-uDYeqmLS)dd{l4 z*s);``HLp|KPwpY4N$^nobG!Z==>$Wv*zh<@zs?;;EfcI@7Mh*^w;$#r?|eX1cG0X32F5=%9_uZ#y>r}22e-k#flHi$cjsq#f zOx@N;tnMx+=$u8f;vQIMB!vsfa|Zby)FxT$m*tiQ?eEgtfvko~h8#yXFsPpYInfe1 zBGq$)vggXhnn_3=_Q+9Hb}CP7g4K6R%V|rIlf%15hk)?#_pjXtXYui#%an-~fz||< zC>D=8DXPu|Q~#^isJ25P9@F(vz!-$}hWQ42mpaTI4nChe%aQmkB^lbt7L_Dz_&x)O z;_!Vvl|JRdtCI-SeWCjr`l;_Rsk2SZ<*4@dm;$v)G(1~ygV8w?j&um1gbrcUc7{@( zgFc|AxihRV*`ZQwMj5Rh+uPf9kpV!!#s3OM1Zr2>b*P%tqwq3SlIE9Hl$X&ZJZdFq z0Wl_<-6>ooqO56?wJq>YXugQt{0ul5`UcZTb9B!M++cXhLeoq<-YQ?m!qDZm384E@ zHFbO3cL9|+sXGq|@^LSF^hoSrvX^BW2>E38dt;ZSS#QE6FbH|y_7(dT9Ac{`<_@=I@R-TA}aLxQxB zHop^&e%IGILk5j<8Hj)&BA4VYj-7ekw}LMpR?LUt&5}er00D;4>XH)BhqJ~nk~L2$ zK1RV3vw7VGT>j3dY;r$niw6H!oLg-e$ysmIxnDO@(B`o}YuKyz_}E99VPmXtM_?S6 z4CjplMjP3SM~O!3K4d7M?cNN%Z7v{4QLz~d1x zy|>&1jgZzq%6YkEN*vLe^?5(JWIQ$ES=(cX#Fx!p$!@~Y{N9Z>#Bz2hq$9d7jy+nI zp%$@m$@%V)oJ?G9>o?i}1!B^@D(Ud`UAp12G#}2`$2fYXGE9uSeCYrse-zv-pg-aM z)cxc_0|*(Ol2!~P@)qN#7hfo-&kh^2F8dM`Nr17JY}h07o?rEDPh|DRVgQY=z5aYz zuYoufV>M^(3F2!DGUx@XnkRyfcDq^36f6)ViIyEqs%d0`Oh-Mxdhc5nX)Z#9m zzr0@YJnYEMezpIJg}^n!pH*t8Wyz2EW{6g!4GWCUcawZfkhxy?6NtVoljQTR!h*vk z_ovvZBq(sf;-m+(dMMa|c;+org*T^|+t%0nAvIyO!Zzo`SD)}>f)uMzOCWI^Vi#jy z;D1|k=2R)rIv0v`U;ksm6V!8m&D@He`Pi@8%n&f{_};bxWn8zLaT$a;d30_T4S2eS zz~Oc8_tpZva~-=dK{9=1kX19y(5CpnV^ie1~-gbq-PjR~{HuiFwx!{+WH|y6QY?AXD2uD<+Sn;*BU5UbF^(dJdib z#h;AQDg5IT7FpxpdQvmS1)KpNVACzySX*GG0Z2;HT<3!4MFAJ0BV3<;_+;q9Fb#n1 z!Vt1Ga#R~y5G$0ORrT-45A~|NR$oP-IGxQ)=xp;bpYJ+Gexsq=mzrJ^0{Y#@niAyugGkxexta# zD*Tnrb>}D;?GbI7!@gSv4siRYRxw35$zpdGJI^q*=v5?=k_HDa0k&g)2xEve5@26{bAF|L{vSsBQes5%eaXV?xLKUr zw8|}hNL)&=6V&T!sn*k_!sf^GoZ{Z-3* zTwGOhzPn9K8yXu^M2qfgMX3Mp9(yPKuO*i{Vs;&M`CyoHDg?qkQ2Y&zX6faNV2e2C z7~xi_fA_8o_Go7LQ8KkSHG4cuU*q|@;pq3suLQSzWX6dsYf;BGCH$nA(rocz^?{KS ze*?2F47$@`A!y;XL2Zu|y0zC+#fqbwzKT+O$xe|+{%yzapoZ`FxN)1hYOg|QHn6a1 zWb=r0tN+bO@UlYeVz2YK9fbN^6vc0KLWhKpzEd3zGUL(l#vEw9n@U{)ki z6uZp@-RTINKo!3Pk%DsHHroXDlTm_}TWhC0fv7r8^O9LQZ;wfjPvTCYaT8SON&~%s z;BDMxIY;A=kGQP_Lgt1c{?88jRDT*?xJ|G1cOwSKPEx_LuGNLa|1CDGthx+R)QK)G z9aELOr~S*>hE-wCSNEdtjC?Rys>IVwZbRV(b+wH8+kI>P1?Td6?vw5t(^jkwv=jTQ zRFHs(7}O066cr~U)j@G)W$$jVjUZaiC!|#HJ*+_Pe9G?3@6!&us-TJ(F!ev)rj*%w zEJ5FSy@aKj$r#t-&QWC0&X_j*l26$!-%;A=aydWB*ur*Aa%l|e;-q=rG%Pq>{=->L z)%%uJFNfCtNanU`3&&aVu8zxwkE_@Maz{3V9xARW>DVz=dQv0hr1FJ#1i84C|X}2*|f#NH5 zF%Ga687~%AGSefDjm)ivf$u2)lV1sYP-#_mX#r1c8VDfGuB;t~L8}?Q*0T5q*ahgw zdm#FWKHA?zHBOBXtn9V}Q5*b?q=xS>jcE9pjh&cAq*-1lWb>+5=?XIq(({NEzj%&X z4|eSczBT-|4x3Hv<>*y~F68;r*SBd;ljL`=fM18_Cm=>|ndx*G)~hHe>1cvdyg-l}-v#RY9(6pKJa5*6 zv_oPsnY8@+N2Ry-8v58UJ@mlL<&2p! zPq3PfVouSas9m%&YPGgv$`^;6#$LXc940pk`PUeCGT-&}5r{hXKK)B->~$b7M7tZ) z(rM|jz8YkxsejZc;tyMh%!s0{GD-A!6p5kyxw^0sdOBZaH4IgPvAhaaFwoyldn%`7~WG%C_5ReNy)pB2k_%$rKQm{_}PX$SR=o%b(Gsg*=P$2j+CkjqSJY z0w{Hs*tbHG<4+FX=1r}@{kSW{`fqRr?8PH+LqE=D4!;gd=KRg?Stw0QI<|^cA~o#L z_U_dsGZm{%_SeGMO+YqKZFh=J$Sr2&uXMGlmaRtiz z7x+lDU**cT5W%xnjDIM^Mz(qQFI=5(K?Cy<_{?0Yv@b&TCqXB_{Z97JfEOF~I_L6T z2E0NZFDFhESl%hll)BL_cqhJ*lK=3LDfw5AADTth%AVy&{T~198^6!bUQU_6U<_|j zTDe7nF%U~byg6D_MN7HsA4kYvwH%==iUY09%{4abjLYEXx!8b%?L#r^BD1aTi0zc& z-_w7TSXjW%@E-zVO^CA%dVhW}5Ht$aOY!tS`m0JLmNQt~`|g)sr(DX%8>Uq~tYOai z*)yW82&_ev{Nq2P7R3r~tsj?mn?n9Ws!pbc6>+_o>)5UPuz{9>Sr=Dd9>>C1<2`n&|HrmaXR40 z(C ze6JBZ2#W6kcDTl#PK-6*(0KToKh8kzC7H{BuZ#Ph0h7vogGCHwdUpF{C%{}FLO>w) z9}MZqY1qer&sci~|)@yYtv|{SK!l^=(Yp&oH*6kgbo?pKh5grJ=ln>?f!Z z{o);)KG2kb- zi{P^4yo;&)Aa^DBsZoKRCile?h*LZ$Abd)52P}nk%*{mjGc4|OP4>KdM^|RgaqVcj z&C6k}5h|YI5d1_lgRnvBM~40FTF1$U<62u)VcJ^-ZH49Pdhk=X-_i#*U=TL@xqMp*uDMSV*jZmH4-7gm0xH4)R-|@X=ud%l79gHkVC8GDCZpz6gUf=LsyH24cO#wnxXZM^epGgy{ zMJ9Wl_bWNV@1SQ4s~;7$vX}!_r&VxaqGwAkzBM%GEQQ{cd?_Ul!+`GGrAe3o-OHb0BxkuB#7y6He9`5AJn+)R^aH5SOedU*M-}+-%nRE*XP6 zkeb!A!miO~N$U{i>kyq=?~9CapT}?2r1VGOg!3vwSUI$|unHlIyM{1yrG&yKE|L|! zZzTosFxSWZsA(srd7YJo7IV_^E4^`+TK)`EXcb4K4i~QZ&gk@oNTx(|oUcC`&r7sL zBPz?kXMjj=0Q{YkkxrxKZAQbZpHEQ#uEvW|alI-|J9biXLAdqN04N9xQ&6D)T#r|~ zbFe?qk}Xg^Z5I+4b-*x!+LJ&LPEdJ3g<0wuV8i!g`N()k1UAGgj_Ef@B}#nLhnvSh zq(KpFY(K=F2 zq`W{5p5P>s1waIu6J4-aTngm~9%1y${!U92*d@ zwlJ82!d-;gFn-3dcI+m5Lv+$p6qS-gBI*{^FnD)hU)U{}qR8)W+%VgVyC;J1iHLG% zULEdzWbTeKj8@Z}&R{P?F|L{$(B@jq6h)1657q&#{U=r7ZJnDgl#SW%x8mqh{}t`1 zu1+x7QqjMf99GR=^J<|@oh_nH=v~Te5mbqA%Jan2%8F??vwa2Zak}i3b+9rUB6|p# zQcxpK`R%umzQNL9eYgLXsW~S2N2j0WMZBkvZH&Tld>jquE+9xR@M8h-#0@tkPp$sD zE&HF>givBX5GnGq2zG7DZeXlJ!iQ?DIQRrWOn{BrhKGeI>o$fvhh<{hc|Ue>35`stQ$_vbzkr6tidX0P4!qEok3Q=b6? zHD}riBWIel+y>(pCrvcRy*Ci66hO+K86VbKuXlq@PmiV0W5*T@d)7Nl zO`h!vY=F>8=h^bf>V;cf4z*_2?>668$HYOdy{&h5`Neb=c8Mgo$INPjNUR?OO~H9` zxxIT+u)SUq;grt^dWHSQ*<(A=O%;F{dFq?(w}ktH291CN#|!iK*TR@DM{Pt}sRVmF zrTA|S3P|L&YFVE3JVVm@XxTwJ`SdSZKP*uU+vO-P0i++ln?P2!DSWJRVbsdcEwL1t zs2`u_y8gdF5loj(`9jit5%`zSsO)z{!+W|=685IG-75=wF-z#kv@3L+}4WMRUyOmq)65Wzm=yH?1ovkR>(A^)2>H#l; zFdbR`TO2lxNAY}2X|BJ)R9aw0|_Xe;hQ+ zZ0#)Y^Fl*<51x7#HPJ-3+Aqiy-$L|Yn${e>@!+%#7-l8Fj}{tby~Aj&d9-O$KseW^ zUJZE-#d5L|QyX09NBv-#>)oYCKMVhD@iQVwqt^w!X&f^~YZ?h8xs0ID%>g&AO&=r& z&Gn8jji{Gy2@Mj)PAaws;-v)%{0z<`SzHi{%Ml`V{`B9XF+@{K-;icv^1y9X1z!CB zKaeeoiA3}@Cr3A%(Fcn4ZyO3|Cb28@-HOvcYjusI<{>-(NgqmbZUiMN-4nrc38Jfx zU?2bY>W(^1l`uAK91m#4%Nw6P%s5>e6QCb0Jb`Pk`#TY8glRjgBHpW|N^nBcZ$#J! z81jk8SED!RRdDePeWNaI@RfK)UW}U5kl)dBf;=OFMrIvJyhVTz5jWZ^OlU;LNE7@ z>}Fnk#F-{j^-rD@;pano50!01dzCK+56~D8L?L1nr0>#i77r%*s6(7trt`Hu%2Ser z8M=UZ9=JPagA9%N8sSjGE4ItSypsHxY}<@clzQ1FLG_N>^n!N)5@J_ld6BHixXF(E z`Gcthl&EuyJh7>}_FqW|{{>v9rO`omB|0NIPt=nQ)Jv{sIWMGZs9gkPt+53WpIwS?Y{wKsTe+QL=w7yI zhH}i+^P%P^d_#;O`#nPhjT`+3^#s3u2tecp5_Sys(>On4k*s|l)JwAKU0PopDH`h~ z!SVP7IL|ADi$u-@{^`TUtdhX0>MZ}1xcE73R>OHuY|ErDXHEIOi6Z>h>uq2LV66A^ z32@-`L5NVttAg{a{M^dp`y@VZ@*2IyCIQmoH}g=)v1|5A)MsWOWw?-V2KfUe$_OPU zy=kON(%8s#CdA%uTLVk~FLJYoScOIGt?3$zE5wmn{__RUI9lxrH%BE~#R1$@ohf3j zVeB}*?!Yod*~lkLuarN2rPkadCEO#WBOC$9bT)_fQf9j>5^4T3Sx z8rnQBVt%VdC$0^dUNEAxE`irY8w9vZ)&mvs(!bH#J6tuwym5ZVjxs$py;0(VdiQP- ziHjfYaNik7^M9EH0Bx-eUgI~8%N2fXVsQ3-@%&zBSXbRz|G6ploreBvtG@FUNnwtQ zXFF4(9;g9S&(4=4U;@fb2DP+s^r3ecuV11Ag@$9NF-p9rs&V?#Q4VSodQ`%Ri4LDg zH?U_0N9VP0zWe-Sgwo8ldN~@CY^%+?HgRUeO~vyky{k5VCTN;BVUJh4U8RTZf~KuB zPRcnJXwql1-r3%9(?UTxc?sHAf`)avm*P{&yv$f9wtY=s+rr;Q{?5g$@U@{8U(Ovf8zb*iX>cS${j zfEVhg!;*=@zqsJLs_2sjcNz&qA&2{BN1IacAVe&{c#uN`zmi^`kUyQ#2pxp=1eG(S zo{~UbFDh^G1@+D$9yduerT~vliGD>)&-htl(PfXc{Of0MX+X~p^b|;UfJBFXOAFm= z1r4PNZ-$2X%L9Oc7|j)_JnCwSPuM8c z?+_Jm`zNq)UU@w)oTa#avMLGmM(||bW*JR|bZV*ChObwh z#kCvqvf4eTZHN~`9Y>+kqDVbm3jWfD+&sZ7P)KMLIgID zV923v!eH!2i~b@|h`tm(AK$x8l7$XXTAs|;u05Re9$Uc5m2lY=>Z#K8iv)<9(pXbL zk|IqT`bak?80aGEX97y`M<%?I(fMBk+4pXM%hqcDx!M^#0EfKMOvK6|?%e(8Hdo|@ z3-hp)4C_u*Q1CWzl6c2B@Yy|a!KUmrIPyCc)}}TCu!jWh0?p-Eo=Bz+JCM!jirg(hsx7JS=%wL@3`4|`8Q4dUYJw%_9$$dbq;7vE!4FIe zaaalC-KkTcx4UT!qEAe_gCAQGJu+Wjhl-M@1Ujguz zq8GAi9r~XCJ*Y~R=>m$DT4lv4Z`T#?}XeeB5_~ql_`;Om z#wuF7vgBx_Wv8(V_6eSNbc#qnh|Lp1W zy(H`$4%B;j)i*0(I*r}^c(zjNWgYH?fPDCz))>QV zfuCd7eY~Xz(n$l=Gz4HAOQi%ZEF_0|mULhfm9f(*w__kd7=t4taj$>745j$RF^A_( z#j1wzYpM7>`3pdEOWw^?)+s+erAK+FA$W8P4g%k|b(Cd2<0o68{Dwbs+!(ltx)Q# zA8Hzk|L#=-*Fc(T-RFu9EU(GDo7b5>&bRq)#uduylJ%D)L+?@Jh?`K!K@b$~rsU2U zt#?U=;bbwQm^L8 z0Hj7f=7R$oLbvq%n!_^x7b+Y+NNO1Qc+Uf@=GUgnRwTKbNi_rGB!~(mzuR2EYE9m+ z7DKvQb+jozU5EVnA~JI=Uq`z1E=7(wJZPYe%t5D@x0Id>RDE>|$$WRJXHIykm#@IF z%k9~ln6man=^h7KvhhzXZ(-MCEQ@c<=Pw@*@h7N2Hn)Y{o3b-%#ljZ0{Fi>8-6C*d z`wh!>$5)jztEa6S>~~L3S2%?_1^ULpSz#udl4fJSF$IVKu4Xw`LZtBHbAh~+;HMX0 zgQ!>{3xp3+tchfzO5>qi50rA#p>w(s)GBEHXcJ0ECB1p$-f>zclCoo90te!$)n3b% zgJhx6l&gey9l>{=lBZFw^azvDjWGqF&Lt2EtxgybeVMqd6#kNklcSuBRdYsp-6R-J zyqh+u0(=h~!?el(3bS7WM37yJW{PXfwXft_mIG{~MULIfk2n(Hx( zo%m?j+YbiX6*?%?V3o>v05|rTIUqUOMI*+8ay2Z;SBr)=xPa26ned=SErdtRHCP~|>G0ec;LiYEi&idQmtmY2-kJ^2x7$g;V36`ECuxP}Z6 zdJ*tVcT>aqwSsos!>i_~l6LFK@5Th0UC_+qoJl&k!8#jIMa#Uj(+W!qzwDRWieNu* zg}Ie;s6E!~!eb$2EUgL(5QixSKnEKH^5y2na9SQTs9`oqpli->`5asXzowJ<=-^dq z=%Hc&`Md9;e!mZGsE~M@z0+!}{q`o1%t}Wr0D6o@C0Bce{ZvZ7lUk3^^a+w$5uEZ; zd2PAX;?_!7vF5Un4) zn&xHy-|_M)(`cRmw$$JI#(pYF$@=a|FH8q0#EAG*oh*loCAG`U?I&;@Y>A@Dz? zuKOQFA_M+o)G9QW1PSAgaCDBp2HP)bZh(Ys{&d)iq#B~$zT6?z;@dxE-~yt375q8f z>Dg&4p$Ej=9?R2OHk-s7y{ti{lRcJ(<74Uvw-`0p6^2>E{;_fA3DkWt4Z{CU;!KTS zepaAoksr8ZhNlOZ3~w~4iUkZ+4re+BU<&w%(jG!MHWLm>r>kdfC3#4@6@c+A6-k74 z(6vLzUTSC|+<@Zkq1p2T=ZQYt>j!cbng=XypnP<3!H~Qd%^P;3{As_4pBTjX(r!5& zYtNP6y|CZ2KrZ#x<9F#BAl$Tay}u+dUA__82fVQ11&bAun3KayKY=D-9B!e+8^kgH z(gV^249%%nodhG_?mnp9kvWz2Q9jrxS~3X(`}C~SY0|P52I5v2uJt!J!)?0$4l`_cCbA|JSL|)w8YQ99O(-+M*FBMBT&SFO<;fs>6 z$vqL)doB_fsNajD;5b7%5_t8*HM*W0Zh%XHoa$L(;@sBFp<&?HdlWeZxh9|t7LxCF?`i39e~j(usjK{|ARBy!3?=`@iU$|td<5>zIN~lwTHYnh9d5PJ33W-|-J&CqY^h%ZuU0k>z6(C9M`>c%9 zuG#kCEE%J`HS$Tqm-rQO^T@qMGjE(|9-J*J$_tSHLqeTFaqQWVL9v(1IImYEX=4O@ z+}@s)p1QpDVFz@RQi;$ZP?5aC*^0|B_~5E|Pa~l8jAJHOCg4Qo>>`ZLRQ#O@e|p3F5N34#-D6!HN<6#k0FYkg#VSF_R0ofzA6!3BA<(bc+ot=yIVFxo^MBWtY z*Iw&aV{pYAepdqR{h4`#fa+ZZ&v@d9s(mS%ySjVB2!NRatUq{6`*^* zX_Kb6@+VQ8=sD9t(*6hk%6BT;-(9l@UAYRS>WO6|@)!K8#Yi(F@IP6UydL-x3q~uL zH#|K$!Ltf%XAJGNU8oI_geq%`z0A~#(T$2xO&=_n7CMF7l>Q*2roWi$fwM}aO&sFB zFpJ3r2q^_IVDM`BYVX+4XHi<_nTm6=0ry%+A~FYu-E-pPib7uu^e6rr6}_zBirdLz zV0SNciw~d~_k}RRWHy0l#DQIqXlCBp2Z8UDv7fVD+uQ~m*SDT#>^HpZ4Wq8GG`VZ{ zgmAXU7qO@K98cAKI~nDy)E}*RudQM$LvyW`I-~^X!{0Wm!qYDqyrR6ND?#he)xKY> zh_FpSvCn>dcC01;r+_e#zAcU=bVTQ?=$~a)JB!(W^X)#|XfU-3F4q0;x4Csy#3w_< z2j=5|n{y$#D&afULt3^A6E(m&2O$DVMK6TSPC=)tqzCexi8a|1`y5gin=K50Ef?oT6pv?I$op_bpkI{wK>!O~r zW#Y(Ngf|3^-uSZ0K>Q+W8ITFp){WvB1r9Nuj28xk;Ke+xlVq(#;Xf^f3MHcoL+;j# z&K-5af!g#G@3Vm`r^?@ScCkJM`@F;$*a&+JkE?Iv5yH7nVGHdHCzv9A`p%vJtY^`C zXMoS4dq;`tI@H`Y&4crxQ?bzhv^h4w5|@u)X9)1MBDrYLvExGRNnM>R^6(BZr-T!h z$NL3_9=_4fvCc#Pg&_}E zxgJQ>3Vf!`y_-5!6L1#+nIJLx<*Q8tZ+6+5B;#GZqRfLOwXuy0bYQ^({*Ll!X!Hys zGj-+7?>l^t{ym@?!88rGaZWw%-vS(TP1IQAVr^cn8lLtiBI5&#JXYW-`h>n=Vkw?lgBGe^ZIMuF&kOsz%15t)mO% zWeEp48W4hA9BuUno|oY`JFW9_=$f{as#-sJ&d*7esuBg$$+*)K?8C*Q4i{9r?h-+V=6K{}h4zrStb1zRSE$RCCgQdIo6Dnb+{m*{OT?z4O38TA!h)CLdPvrkKb6sOGps zmEdHYHC(gl)u$NfeXKGvDf17(G8vt{{|OtCi4z(q>*UXYhk4XP#NOI|t#Ugkvu z_R$?m5qwS0auwqYfTu!eLSAN&nBFna)`$T`G`wmlQ;-t(ey)kz-y^qUv`w~{%&@D* zG0V_c-`HaRkKFSvcy`o{Ys1ra=uovK$RR}~tiA1!_D$D6I`X9E7`4+)ZYtBgA9D>% zwwW6`UzIbovd4{HHmO&}z}lbffH(UJ-KTJs*9zm9wvlM0%+QZKji97^F6Z`|h`FJO z)y(@Qu1^EVZ4RxnH`$6%1AFvO($~qAamCiKdfz^lNIv7UsbHjauUNI66wcv$5^21@ z2k8sM($$odtIXQ$J42fudDa*K_qZ(ufFyCEr&hf6n)P`9F_5shiGNOl?5x#Ly&=-5Qx;>^&SGp^avEILk3^Jz(h{|IESh!vcfTea!9gp^PSZDdKi@8 zZYjLOcUg5y(o>Bf8=&x%-Lk8R zf2aN^t$fECTyNKJ5r2*3d$)lIzS5U9GU)3joAFR9phvv&G?2UZe08RhZl9PP<+ecM z=ekPQ&3yu#V7y~k|E#Q>S>6k+%UJ0Fy=|cJnrIwm`;tM{F6Be@G3I|Umo=0s`dL_| zbU&zhgu+16C^Rb%9c_K*LEn71aubrkxZmmi@?^yue%>w~nD`n*f$HUDH1)}cv>Wx@ zRpncm;Qe3dav^!p#cveJ?yI6Hn|A!6G?M_0QUdWv2s+^VzYfVvdwlbL*`(|NxQ3C@uS{9q6HdwNhi7-!t?(RN=$4=dtup5k>*uuWI~9`eP^E8mH=39d||=%xr)xR~K6Dhp|alQ)S7^+H%Z3o$06& z&}Q^TPj$a^8(=alHxrrv+b0ks8|;@wBXqMlfqPHrbkFyKroN$;m*Z)g{ujsb_s>*5 z+sr^&eJ{cdW2Z!;JGvO{MP>qN1NmPDAvvB3;4Qq!fJ`vTC5oZykC3!m3o11ixkTLY zIm*T@=~b59N#Q~T(XcDi&BZEnec?_#CjtSM%{B}xVwlUYID2Hw2sILEVWxznjR9N5 zrSL+CN=mV<-ttnm{iqF54K2>f?g66pA4lIu&=8{``K=GE6*pu+A#GR@U4UBEO)P|Q)z9-O7ExiE zd1WchCz{w6XTskQHZTZ@r*Sk+4F0_?(&(2iUMPUqw!+MyeSToXU`r<8tmwHB9d``p zZYmM~zQYT3@aWn0Em-v@m@dg18v)zyy&=ttllqBUbbXWF3tmnJ7FoZb8j<`>iKs+y zGd`oyKz__pM}R`(hZ51}@sh&mJ&}vaGc`N-TtgynW_8+vEmd3vn@`dcbmwmX<-`rV zOGqf<3ZW8;z7hpOl*O4~Kfh2x)NRGBev7C>;|3!R~J5-YOXEuN;&8p*3* z)TkVePAfiNJ$|*Xln={Wc(MlH>#u!P%qNDaks^$Dm(E$W__$G8ksG%JZ%P%A^fpBQ z5l<`6{8&;QBA*X!TpkvanQyT?+^YAOl!4TY{00s5R@VR@7KM3L1=DdAgqX%>FAu*9 z7@+QXV?bkC&;17f0 z_o~q*HLTEgnHCwgD($lKo!cwDeMHA+>yT$1(nq{dNgyJ~$=sx_>cR(qmXF1UzT{hm z>EIInd!BHM@TtKakK)%2#~9x0FYFUyOX+K~8NBl_#r zzft!VX8u&A_Hy`tShul;49_t}TgDI|ucKR@Acu^~XP4nau{X}f+hAxEgw15@PupMg zX~A3!@pOB4xCO5P3GoN9*86C!P$;_lKXni}OBb-cp8a6M_jG{CqJ5avEEebBd2O!A zvc0!GFWyVPL4`<#JH_|V8)MhrYRU3^c2h3>a0>0fNT)iy3r2T6mp)Sen#CJEI1fHr zf9R_d-n;(*d}{G-d#I7~X>+04@!dq^7)a8(r|8gIPdowB_b#iG~M34ORI1N9Fqv zKdiH;2yT34CoiIGafR{sF9b4x+mdm{SX9srj<8w-v+S#CvJDFXHD7-3{9izmyC>ha zc;bjJo;x;HpWNjhw!SPsEWMGmzR+-y(|`$Cf1)3Mk1(2evQx15bqMqy&7-+!cte}s zu%bypdB7&*=FRK`zl+s^^Rbs-vGHU<7CVbPIW^a+^G}0lA+g}vcW?T(r@P@>^8zD4mlL5ZK7Wu!XK)_%M_@7VXTgp4Y``H<`b|u>KR$h-UI;20J*bui{ z?qWxSBy)~<1-uE%w|0bSL?l;~;9Xkcd#$JqO(XKH!W#p|9~X20Ar{Hc-(=zO{g?-DKi89hv)=`TR`}D9 zPurvvH|WaW8(V#u92k$aK4y^oKvGvhX_Bzv`8K}0vcld%&jWo4I(&K+C|XBLrR z1z9Ji8Y)}#)z6PgQJUqT`#pL} zwcn}#7(t;ZEy5)x%+YOQ1=Ws91Xz*`yV|Ex_i?pH{!lADN`cVjx-{TrU4-C8B(@7d z0XeipW)okzfQPd43A4AN$Y{au51_k@ArKo+%uYf=3ZVuA@G_nImKyf_#Kzo%IP{wN z2^4p2T8-zPIFx@E4I)QD)=eeDB1LCe8S=Q`k>;yD<>pE!m>$Vvl1C%WtsHy** z(O{BO>h~2H=<8byeo{;PyLOIx{3hmej*wrE!j}`idDBeL#SUZR&N0r=np&914bcP6k3pGTEY6B zYV|EK1m}KUcv;ra-pHS?lG6$LXHTJm{5u=hcUpeYnMbpLGufWECNc6Rsss5O<6K2= zFchp0y*Sc6)$sPx^Udt^nA8=1Bd$_zP*OE&UPgt4MN$|*&#MIN#iiKUr7q&is|EY* zm5!?x-*4gv%&Za(&74o14n%Y$sup@g%ucT+2^J`_FO88wsnO;BXmaL3JaTJiV;2lS z0&B0TP4#Bp+lAG>?Rc}N348;)y)Nh9!PZ4RjRmgXdYghft|s%?Jvs%^{!y8GTDhb9 zFTU=Tef2OPsyUc_fRKmnzv&4(1Y-Cco}Sy@xnP>BAYvQZmg(s2R3&Xjz@%<0thUMY z8Ent^)G_;1uOZ_z+fd(+l4Y!#FRun+smEGB#I7n5uZ7%!#yAVpJ+!v~a<2h|P&OwL ziL>GQ_BU-jc=de?sWP{@r4a%U3z_s!|s=PciPOtfo!7*x|hlpzG(r8%yVrLHn> z7JUN8>dU0d>?CAFRgMm}_Qusm2Y!(;W^xQrKP2rE11IQbZa)g7PAN!)_wfXmkE7{b zYDWsMvkV8M@6iqnjUyHCe7Z91#kuMsZXPsLj7Xdmc01fgxTBcaks`wnB|AVKUPepc z>wNQh7#0FRUh#nnqyph=y#?IXH4?d+Cu103yAuBBvdRu1u z@HICsFsMJA=iQF={3lq6)9oLeDwX)Nm|BOTR19I&~QJyrX-K42BxM+wwH>aB%rR|LFzQJDMH z#|1C7^8G^hw=vLM&u_awe{mW2HeG67DI<)&(-A9H7Z@FHjBi{*#;08%-?1Iup#{Gk z36dPjNMC38ouAkwI2EIK84q!;kB&Evmo)2tXVoG@SQI&Fmu*r`Kz;fLV-?|Jf4H>r z)r>5gXgf0j4heDa6Cf2y>qS zw~>-aB-6w#tujRm-_w-2MvZA!!s=Yor!jq0f!9$WiegXg=Ai_aPk(nDi%Ka{bb-1#(k z77v@hG8U5c`^VL~;T1Isq3Mf+N9?SeD6ES-y3fY{A&$$>{Djr(Up2D$LDKav&*w^w=AlLU zEIicu1L^5v9I~1s%CQ3Vm;p#oJ^G3ic^WvPOLIlqf?%c#qu?e@A*E633N_QZ<%{?9 zR{v}uRptu4C((bs`P2tRJCnM{dIEC8^?LNNvRFmwbIuMT>T{#JeL~Mh*|$%_+$7(UN2lb$cQF9z>lBl-Ofyd01bP3Z$gXvY-Qw-yeeQpUo)%&I;NfP&NjBHt0^tA<3rZOWf zf#lEblRvp}FtBZLJ$zhgnX!Kbou#ZZmM#Z7Lhaj&7B*Vo5V^-1m<{&pbhoABm94F2zIAo0w2hX4~7PGO8&HqBapwf_qFugttlt3Fy-FY>l^343QLe{NL?wX+s!Pd7-ea?qVKTR}^Z4-;Y)G2L(&e8N8h_DQ~KI z=|fI^1Lk>AUx*e;oVZ947g!p6-(zq8kVdImxcy_IU^T$HW|KGr8awK~E`%54k+I%_t)+}> z+fYB(_gj}xIs8_yDJ-RHOf*?U_p*_Qy(TR|iI{~UGGB6{+$vOa1OL~$5XpZE`gWd% z#MWJTwC2KjTYx$0U#w;#Sp~=ndx5`n^{83zB0JR>oX68r_OA%PD9+-DBqG~(hliSk zuD$xQhNX$$hWK;*dhjhW6*kqAhirX7m1^bY{9Vi^^9yr7i6Y*5 z#4&g*V&zR2RX%IEh(I32Onx36ltmbIaY$MH0Osuu8>p%Du2EQX^){-0p-uSGQ)ykr z?JKSDI{I!0SdMxu{?FU<&t;Zw5#*VswtM};PIcfeg;f^P#MG)ZAZFtGD^qlHM!?4z zV*GF-S(B}?c-(@ce-TB|2#LXCI?&c&rME5peyDsfMuea~jWfYLq==4TX{&xcd`w!ZDu=EGjnUUTM zNET*V@>s~44ZeAAJ_}s&|5)MzA*Nb$t;mhQ#vc=y-H`bwZh=!gklD5iPZ;jPOmxP$ z(C~|R3$|l+Y$uRCe`gyR1s9`nIW%yd9sYONs>)0;`cJyqE8)Rm_k}#b{^`fwz5| z{DhKWWYLLZ8(5g}lqPvTHVD$0KKR5Aj2n!#OT+wMqWp!@2Nrx$&}ad9^_rn-x^N5& zVHHz%(GA--=L5a#gTZ-OxiefKhW~0S>};mCr^>uFDzXMULhP=NCW5#QpC|LOn|9}Qj{1Le}q7jt;a@88R}| zY|pQHuJxK0Q7nNX;!@R_ehNt*$x6)|YVk=EYl0sQqAx z<{ZEGGOKRpdRNHTb+yA>K`X=8sVvD_Gj2V1B7{K}8Y26ZhB~?@p$thBphE+j0^3v= zfPiubzT>4Sxm8E@;-`=dUE6KLDj^G| zmg*h?pD)(C$;4`D=bL*F4zFU}tnkVVrxGfEc~@rkZ=H`fzzP_?j9)jhxk!lo{=6Cq z3b~PPM>}B5{qWz*?rO*^Ej6ySG-nY1P~oSsL$kNhd=R4nbCI)~e09K-Y=hIUUW}J4 ze#ZANax^7;=9$hO%`ErY6UkNHxvDprvB?Z>VI(Ya>hgD+VI#FAos#y^Nj5B3B1{E^x677SpYIFHIjV`W?M^NUArYEsW z=HYmx<1c4F0W(9!g$#x#*D_MLWvHG9z@HqLR0hI>$o0w4)y_leHNYJ;Prpfj3#|L< zSRMZeBvLNDAO@mf_#ht|t_NoSxj|ZdHksqzarqDF1BgK77WlMl)bg%?qU3$zv`f&x zVK48|&_n(E1MbggO})cST52;El2*u^Bg!Vsgx{np zI#EC<^3Xs8KF)owH=r71RIZ!dJSn5~exa_`X{pO6SMKJ{!LZB1?|x?qnAI8dv6PB3 zSICN=RQ@W>GNoz3W06f(f-vxN@b@re&{C&sGyKE`HS!BdsCO@9*`gmy!V<}et9e|a zIbEbmZDNc(2VO*I@2CEHQ;W{)GsPuMR|hJnJg+}hb1<$qGhVpH=^UyVo~lP72$#$} zX`k3625C+I9er!0r_{kJVEoE;rjf`a1O5m?8Okk9v!;$apesuOr}1p+x=C3x1n1Mg z)s~e_mZ77nZM5^pr}A*R)1~I}i$yHA`K4BXXh_3VCQi00n=DdU|GAYmD4R@^`*>Uv zRdnLH8lJ4$d;>>%#^ZmE$jbdAc=Pqpxl7*0wY>>Kop-YZ_fuuzVt?MiNt=YWMg? zM)Zq4e-}{C(-)Wi)ar>x9>`ym621GLMtua!zz!XM^7QpJwhcMyf>tW%<)N;SEQcTI z{!GEocmzHaSO|7Nn*jgmQoNNBsdA{bXAj?^abI(8SL7FWwt)^#lEI%AS=pJqW{(R1 za3W)RT=F2hX*u|^C|EL-H5)~HDnEHaW=D0tH>|`Gd(`b+j<_N^9(5`4@@XS6U7gZ5 zLFakn>E6Z)4tZ@y!rvuiRUoNL4vKKB%Ren2Y#m&0YLwl+e)i}qu%@Rh&rW>`%({ox%jDE|LaB# zBJ79V>3vWM@q=ryjn%zxuV)AC#7-2Tzg|ulZi&>P)3;#Gl>YJ2wi6htpx$;93a6rv z6Rzxh-2AF_AVZ*1Grk3j+yF}5+uUo>i)>;nphgng&s~pigo@$~E z7&L@>QdG#a| z1PU8CAKk49BbBGY`X-Z{l(Anu)Fm7HI0TW8C84&kh98!-T>5Z=e(3))hVMKeCKqn- z+`pDE{~ln65x)R+iSpz)0LfI7laEgT_)K`Y6*;YQCkXlzD=Wc%36ix*I0(T;w~LIXnRW~3g!XJtz!gt+GL z+k3jh;IK<^v6r)!Fu_YPlII4~djak-Nj+0RSNccbQ!ojenK4Fhx9bs?E3TjUw|`M8 z)exRvS8~r(%;)cI+9=F5kF{WSsa;nC>e)RCK##9E?!p#nSh+g+2t+*4xl<~kq8t{yitO5V37R~yt!%p|Lq#VT9#eS0xT zM}bP*FY4X12B$D8Pe5D`Ci4O>(?RglEq<%MS8Oo``Yud`WfYRWjf6b`&+fdzW?b$6 zFeV|FJ!y$F;!pFDm()Lc6OzX*d3w!2t2BXkbD01L{Ao7{yj%8VZh(HePZ=rXC+8j@ zl#UMrf^nroF!!~c;Cb}!aT&z(WATA(NFu4Gf6)&km zz&>JLqZ4VRUaE7C7+>sXwUBIG`9zK9INT-UpPP=a@?D}c=$hZz`#??!##n6*eSmdg zB%>)2QusCp6kj6SmYgYNdRXjEijP zqc1h?*KM$1VCUa~(=QJ|g!L#jF;Cp_d1J9k_fU$VePdZUCwhLSFGcGZt5sfOY&utD zCS!yA-@>LS;@nW z_eS<4q?9HED>ynwyAHlt@Ci-Yzx-=i3+nbarw_BBVS-v-@Q_&s@cm^@XPk$vzAnM9 zs<&6N%`1r6WWxtsxn@x{IKp@J?*F$xJOzA{%fw@S%6!{K0f=5J$WjpAiF*N%YOKRY zwg8?yJyGK`Azp;>zDutY(;f8*<(Pux(fb?s#R?LS1Gs2qSKET_+cn7#GRGJTNVW3SywcuA^K+PoLZ zsKfLQBVjexe#1GHTqM7Ro>=0!e@tp*M(--;ZD#+lpCwSLxgqZV;bP<+)soDBLA9hu zwkNwGC(m$PR=S^M;9C~!{ge6qNwF2AO!w!Z|l-|DSoH`o*7QQK6(st3<0Xidc$ML)9nxfm`GO*z|vaC>6 z*F14A&vv||F?M6t-m{TwqXJE+zH|XNCw(XdPqO*n=GRL-zl^Z&6+sU<@rYO$O|g4B zKLp9MZT&FcqrMjM%wkU3Mx*SV=R4qLTqa_aVDMgt<^FUjLzNhDx?s0}qf{CcSrgJo zzBv1<#6u|pYC-L3*g`?6N|D?RIM;Pznd|*t^CYbJUf6mawY%}=3<%S)pyD@9tTx6v z6qw2EwS4jU6X6cJg1t1`jq7TVLyXGb!R4N-5!=IfyLWbf9Tx7U{!PNbWbG4^dI?LydIZOFxHN(2PRJ0h^zt7Qyg19fTwuee2{{_ ziMYm+>0)6zIK0%5AyM0xePl+7+i$J6tPIYVXNfnn$a-hlfi;u)bj2BXYR#F`*cW2^ z>5Js7YaZdbbN2An8Fg4u7 zx=FrthUV%Z0}*h!^fBr<`9czzVdV=qv@Z?iHNgXn|dq?1Oz2EcoGDFYp zP##!Yeu=;^t}5!t_k4S7>nuio`kv_Y0RqFi5`zNu!P*|_a$=!L&PYMD3qz<+h^iDt zro{<|J453tL%qI~Q`Y;;J`{H&{R_f@0+#+hn$r{5q0lQGP0B}{^&gVQKPRcMkZ7C1 zn_w+aCJv1E2|q2zpGL`MdIPU_9srLZH+a%=@(pybF|Uqtx}Cg-ak}gmInJM8n&9ym zS?;6-7G+Q(d$RW%^b2btJI2RX1X8q7inL1Ls1$f}cEm^M3&{*Ag5IcH~NT zt1D}-aPjSHvZ2?UqvrImZ$R=SxgwPR`~Z#o@DkD zF6>UxiV9b^*MgrNj0pVpu32caF3BKtwj%+xqo1U1t{7M6JlNt1hV5~f*AJ)jINrCP zrxy>$RyPf=PNZM`Vt zf)5uKtvqkWEB>ukQ0le^NqvW}E5qEMElXv3kZ=AvH57GC+949`yr_C zOH<7_o22(e|Yn!(`zcG^rf;>YJrhJvuiY0y!;KPCIHdio;c&rOQRJf+W#g)@M|DsI=FQ2w9+-!ps@ z`W3k?#MNu&(`WE;MaDdavI6In3^WkZ6i*|8etGTRK6YW=X@7@NvVShJH87oVv#Pm8 z$v@L4)~%&_aIqmu24*I|*&L#nPb>9W6V`o&UeBrGq!W<)@ODz?CFuy14oDi@;q)~A z?@TYT>=U{*K^O=X|dcUUC&wvRj}qfXfiD;MSb|ZlBaA zN`)OR;1F0P6QkhOcT!usAF=|18}zpFH1m8PPtCp}Kmy;*VolgD8toJP)ch&6 zs@N7C!WLJQn6sP%`^a!F#E#06BsD#+bid1eTBRqw#8~26OO%w(Ujs;PVXB9rfZzO1 zH!ITew0`jzTRL%sW%ty0#)et26dS_`QDM0cFOjqlz2CA?pcSGZ0WgLzqaHD<*d(pj z8!*>BXhgnnaYQ40ycCvnRu-!C$7eqJxk9_ujY)MBqf4XvA{=~4XY>43M>bYdZCPgs zQC53pM2@bV#~8Rs9)amf>+n+6sH$kSk-MQ&P0bfe$al}_uc3+G>gL-ux)?3ACX^}A z0ocDpk|fdnZ~GwKq6t z-sy9xAEp$f;SDf2W<7Ic_asvqFaeJ)GTu zV{QLn4_%7h=$Xr+m0W@$c1)F}AbbO7fy8Lzmsiumfx?+*WhI%R@TXOk1Y)n#jY*Ds#pO$Lh zA~Fe5o!|ecB8rgI$wb2;O~0K15-8i>XXshg78SBqXmB2L#B8vwH>HZWO7Bz5vGy## z>h}~G^V#~HB^QBXjiJdwZq}1=!>e)~zb@?jUVZ74LBBd$l?jLWrF#_yY<^*WsdD$2 zIQRA9eDvq79=ukcnt+cQMN{g1(eo5F#KbY4QiIRP--jgwR=)@#8&vD-oGjOH+*?^O z1hl)D?+ue#rn6YxeG~A(KQhcXh56AilZ*@|FTC@<{L9$>?`%tm_b;pSywRaN#vCkX ztLDi16Ub1Bw3BFIM`?m`L4;`d#-~~i4a-aXM94tw31*ZbwG7N?aP@{SoNwWTh+k7V zgRssh8`YN5Q@3lkBs^f!DB$!RJ5E2Mi_ob`QT;$lc67y zT*+J%qlWMAr1oM)YS**dTQ7KH7&1mEUBN_}_XmU$@qt z4UJ|8&WVg<%q2H4s`PrYt$T&IyIY5g8<`oo$K7IG)D6d#l83#KHKR1tD*X_^Z23`b zsn7tOs3D|!54TDh@#CCe5gv7GTzuE9pSMUOQy?9IVxI+d874;bfKc*YL)_Gzz){4~L!Ml`t`x0=9XWRBMm? zo#}dZ(DVx&u?YRyr8EpCdjPq;=wbPi#=HP32+zJS=D7?x7012+cwQ4{6DW?{X{PpO zz_LWtmDuxOg$7K|zbVOLbX{44_X1SH7i@8-$SN~zl!XVQ2q7-f^~vMD1I^T$;G+2! zO#0uIaP1>d=+d2Q*FEzWhaH={e*efT66&-euB%*OmjR|bW{u3tAEnn-cGZO*(>(t@-kWSWDE$EVSvjdDph7Bfl zelZ>|KkAxEl|o(J6v%ia8!UT^J}W_PNI6OgpJ3iFkyVpsGhx?1oNb?0BAXcXditRN zGj{L)`1d9%AUy70Tfdy+Hie4!(IhYVwx=m_qqh??X)0%avrbY{Kw*v>DNUqbaNKBx zd@#5sj5#@sU`$=cEvOeHA7eL|)0j22DeK&*|GY_e+5qE;R)6liX{G33jpO=%0*+pM`+D)yh+V%j8gSruKh{W=4r&&{d@jka7 znKG&0Xl7a*bnLXIw&~VRz7R4fB8>h4nsoI`y$%;8D1#D#G`9ArCJSOIPA%7AtF*I} zt+rR)VR5_aFD;A)rCB*Ky+5<6;v_ z-!rx#^nDG-FEca@9|+mf^x>Zff0iWeqVO;8-1Y-|o%;3~WWICG!hal%A}4|t+3u6X zs$BCy6XRF9PDJy&*{>q5N!@a&bK;(#9zcA7EF>(zSmsPnE$FjwR0XujQuuIr?&Fp0*c_xW}q!)zztS#rsP9to6|E~y3S}6ESvPEVSi1O(yUHH9e7x(x7S2L2N4_i4kQ=;mBLbl4hf z3J1La$``w=;`_e`V_hL~nwGW)qqcF64?2u*rx-Q2vMh6o&04>ND3^rYd>R8H{_8Q| zrTq{DTiH^p(k=5$&3|;%uhf?Hww~`&%U~Dp*D6skc=FpqO<<(5PXCujDCu1-%zXIl z!+NK;--{j@D%XKhYWCB8zCKVw3KpYmY#?6!8W*tNIJnKk7N1ANVN{=fvC(OZ+?`6o zVW`1QYoIfI6cm|R!+P-`3T<5|)!AjfL5XKrsxipgO_Y$moX$?7!B23*7}FWTphUcm z(w6H?kD>HSPWoX)<*}tpmGuyTZW3J9Jy*tf>ESq1jP4REeYUc!F+G=kt1m@!5k5&* z*+>@NK@)5&lbqC-2Kb}z^ETUX_I?oY*Oq@w2a%@#B)piU3%D+ROK7S1?6_XqR2wn3 zhq_LtTVv?Fo;PpyW4n^$LZQmAjgngrNa^E6ozyA*L7njoRqP!^VOX6ci&E%Q zWk``q3d0R!JX_a7spLs3<%GhpNy6@L@=@EyYNG}uQ>QPZD(YipKQ1`8%{kC;-pH76 z)X;hkbLTcRu9eG#K>RUq5{TCm26y|scSC({%(0cy_(jF*a3Sdg_vP;U6JvVX33T>N z9`Ssod|BrfCk$S$J9RnKC2T%)|J3)x4FQIzK(nR#jP1enF2x)v%^!l)%W* zL!UACZs_pc;|6@>=8(YYFw_Nj(R1zP?BOnm>#7I&19_oG>b?dgur_c(z!+tCvO_rAsFyo+03pz!TZL$wzW8WZQY zyhQp9Bl^D=8j;rTg#dJt$7KgxSYMZekpPtntNYpCLMQuA|I87=*8i`A(#6bKFX>L zDYMrx+gDl~kFI6#ns7er!|95N(}vzeMLM%^TNi{b$CL#Ey+-7PpSYYMoH`RM7+z(u z6hpe_W4-Sd=Nujvug<4f1-%IeXd>+AZ4|6N==5e4AWJ|GE5|x{E&nZ)W=roAO%`n6 zza=^?XiEzF2x3&~s(55}hlN0c+kk_kBX7XuH`8|y*7AWEb^H2UIoELZ3^_>&75T)ZIm({> z>Xi<^$Aje1`={mi4Wm0T_xxY_oTA}s&ljsjWvzqA=oho1`Y@AMlwTzSBh!WuwJBeiY~M)HojZx>d^ZHmk2NS}<@6{^Y^}*4$Zo z%QU<%2d7$S*OJ%2*eqZ=HHNW=Fy#z^G4`9T(^ql=d*Ato>-_i7;?^v^Dbp91zuUYW zh9<6A!uBHfKx%UO%t+31)Up2e+rln9Dm-Lpc>mJ)Or_Nq1ALz#UuA9J9c9}T|KieP z%EaRDR0aPDm0z;6Go%^Anqb4{JV(9mo$$Dq0u@W*$6hyhO;|{@CSg4pI?p|1s!)}FTl(1DWu2! zLlSq?R44kNs@BQ$dA~}a$%Aa)M^c2Wo~Wp6WS$AxAZ@>g+kDhp5n85C>rY$@d?2C-E!=Z z3P#`1>4a(}VvNa=*~_L<3sNqIJ~sgV0*31o40@HM3E?n#mU5dAg*%BeMba@GEY-x> z%*yAhS)bSO1PXRE|Kf&-^fZ9UmYkrY*XgPTZ)eFG)M4+O5#okA%{3ZY@WYO(zw%#G zx)~Mfw-ku0evsSCOJG|si7h55v126k4e_yDC^NAzli+K?c-DYdty!akXtS9-xS#Or zgRzT)kidJA!rZd8>)Zb(YmLCqnOrwr_7l1R_Kh;3}P{Ea=`ajRTW9=Dw#&)(TR8}p5epL z+T=UQU1jRg92ewoBe=h;qA(%YiDC|kfMYZ7`>*oRA3Xe2t4tO~R{Qy~hLf*}f&KO; zX=PGA{(UJyBgMlv*xruWlVN!@btIaT2`VL88Vqr-){Yki8U>j4U5I{vHnZn64MhZ* z*^k^|iN0w~|1i}unvu>!hd?X4^2tZz2uNHzaGJ79!94kH64MTTf8zzyJv*3CyVDt( zFm&)d-*)J#38MH!n)&}JGc8oPZsi1z@z1B0sG5LFQl?TVABebv8umDE$di-x2P-0I2JQjD6gue^!3a4%dsG~$cQW6$NZ@J5eW__RpgJ1Xp zi3i}#5I=>jvp+3v8qN~f@?IZkdWc&-%%m$Ym{;#Qt{KDKi=b7WY}SXXux&5BXlcXu z4QR9Z*G0VKXt-7n5566*{@z5$W(^Ng^!&ftvOdb|NZ;dyIKLO5)ED&zim$y=W~F9=&h8}IxOJyHqh-qDWr1V71z+$Kf3OxFH|;3YXE!a zi?Kj>*W&p(3nWl1RK(yFU9A^TcOZFOS)4a&a+J}K#b9GhO-D3Y<|}UB)fI(~P}HKu zPZeN_A>O__P*-LvZLyfXu=FkgrbT@2OguyqrioO&9ZEmI+b}+iyob>8*vpcKzem_X z5&$IHaHIx~qYq<6nCeZq~4|?tRvjOTT(v^(bqAQJ(%F~CI#7yZj{4<=4*Rl zFmr*VOtm{Q=?rji5k+YIN!QVC3|KNAcW{G=8zx^GlPGmaa6?Xe@Br!5P*#Aa*YY@N z6B#Nsz#pa;-%)lGVxvp2#-t12QT;Bt-+!0hjKx|-jx`jwQs(<;p z3s?0#r*7f2_euQH&{ppMjO$}&>(0I%_m^2aTc_}paxLcykGhqrOs8}WZ5};?n|M_> zBI7KZ4g?ppb@+X$+!J*Qq7pI&gXZHW*Zce*ZFU& zjT01|9IP~k)0ISDvGSeaZcn0blceG(NlJ62F8DzSEpO><$~uWYsa=ikD0WKMIn(`)^aBOwxaS=lJ} z`9A4=Ei3V_r)m5Tb48=WI?D%LMSW<4S;*uhEE$3r7#4Rgg6cA~VYjY$G#|jH%eG_@ z^CA{!Dx7c720dqg@OsUVg}ZWe#-x@heBX@-`+O37Fv=N(G4JY{z>=mQ-nlztx7u{Y+( z=y^Axm8HN}4S`7FUWAJ?;F+14e$&X|S5%WTB47zL2&cXQI?l>$^vi5w;Q1T4{JLh| zB`A|Nh7p8Iqou6fz{`#aC75!?Y4LtgKnUIW#h+=@e?82ib z3lgy4zvpbv3$$))SgsaBW(`WHNx_pv-^w>=IKM0)%i|c>9IeiB+292h~>a4s6tpScR3)e09qh zKpcs5Iy_psr2jdqyN*hG`JC1jUM_V>7Zx|7J6=bM4FQmtK9J!q zB#{Iv^OZN;BgDEjHnkvYmWpa3N%wZ>=(&~<_kUIejMvdGTktlRQFtfzwB)Tu1^mqG zya(wN^oLJ|V?9nrJgXWSp$IP3_94kQNwuC|V<$Wq;vRJhzvC9m3cSZpkY;g)G|;-ZI>R=W+)>B|d`Jlr>W0Cn|@x}}Kk%Q{bQNSmC2{wu(VP zEfZzDq?w)-iM5SwT0SB z;CQ`t0nOr#G5(*cWz}an(3z3Ok=ZA*FokY-x&Q|#A^x<2Km5o}#6+9MOkZfd^_KxE zK+GF`OOs|DD_u7@zlC%(aqbX4O!Ed+88K-JiMvmLvkS}uqP{9e z5`ZitdqrMD=8|9EC>zfJCyOATer1dk9;TH4rZjOKbQ}e>d`KWAos^3gWcZjziwajb z-6!NK$aDbSXpN?rFTP7-BSceFj#pxk+fu=yULXYYp<}7i&bJ?=eFZz9HkbFlQTnA> zojYf<*RSBADkSK*RJ^_j%#4;4O^e)c(v)7R?1z+E<7QKjb5k%A-U!4<$o(ra^nY%V z?za7&(EH$MF&_v#Kszg~80aUc27-&CtA76@ZGf(t$IQ>5`i;=WY;Hf=Q#clKHB zcc+7R=mu?eHH=B}{kxofI>awWDI&1S5K=vign;nwa)q23)}gXK)-$(do>lYJpm(`~ z3+4<$pGy0Em(Yrrr+0it;VHGCmk*Vd_e5~mVkD}GW7tUU`5*tfKFlK3=oKbZT&A%i zqQC7`D&}3=x5ha(8)R-#P=t|AMI~bI?SGc2-VW$TjxW&oHUd<^c<-H^z&OOFVIp$3 zvIAa*s)0LKF+L0eiApD7a_BRnv^WSFLcxqkAN8eERZ@QdN(J&hhU>`qe;0YJl>Pxj zlWK@ve_mdvgZv(V59W*_6^;=`;e%>3H9@KfHAg-4;080r-!d6I$l~__|M%G|#E-zg zP6rcD@_f)xC{GWYam}Uw^PZ95L;Eaz&CY`*N6aIWZ(xquj6sTicO(&~WA7w@hA@;( zGr$d^sFFv$OcI&!H9`6Nb<%h*9Wr0`=L7Qj$Rb%}tKmdbcw@ zGFuU4MEhthGU=FLTOZf>J}l8#K1y4LIca?pQ7$|ua&C@xXU2---@dvndt>ISX&X8D zn|~m8Dq}Dhrw#! zcWP>$tRnpqOZ$$>os!Qy7>U2|MITPy59W7kv0&plGu(+w(S(^=*IGt+-8Ud(;Ih=@ z*kpVXCETU)Di{lRVF~P}LQO^t_Pj(*_M_sY)l%iQVyoK%QI9TvgETDdc1p)B6!%7Q zU9e9;2!i)tnCo)M7#*97L~-ApBA1valuz1v>|T9RK~Rexb+~b8+fZ9L^Z#c7==Eir|7fDu`?rE6oyMHxF|Sjm9Bmf6E*xxS zLy6oB_a&PDa?B(ko1*1bL;Y?feAgxQ&F^nJ9kt8i`fsPoKtBy{cSGC_v($OS0Uzn1 zg_w82K_dgh^M?;Fp+zouXAFK%{XI9sch}7ihk(7jc8u2UO0z>iRzB4-BBcTC?bnm>7J9GR z-_ag;-5A+YSs7cbt@U?}g~dxwJ6`V$&w8!D$dV5^Z73fRXKdCWl@1kY#p(Zwk^>xR z<6AMMem>%eFORspClT4BAC1bDHZhOiu%SrSi=>4QJ!qMn7r=@m;;e$FW@)6-&8s09 z?SWS_ouY5T4$zO#*HGN9e)50~D+)X9QZD}eHG)Z-Lru0Zt*9Kj2U8<_A6L0jGgZ>^$?hesz+y`XwUMjjb`A04iYMP>?qT z(mb(jS8GE2j6+vLmO;%BJd8=)eh%q-Vm?NOX@fiz-B%GJ30}dR>EgR_7F%%BfIEtv z*;T})7}P-91e1O3_HujD=d%tMrFUGze3|Zm4>Tb zdJBg>;_U5bS&SsHLma)hC1AjpfNueazoKWT-+3^=x!(U6ivk3&0_3MdCXt|T)icMg ziU#c=Pi}vH^diiE-`M<}Fpwl(@SNA0@%{)D?Y?ik@vh|E6k%7)(fv>2oFPEjv=xhlv&SzU~6T&sv)O??3TS<_I_>cHQl$b5g`PXmyo)vTYA7tCI9>ycyeH^KGB{$rM37=+DToDkMlph!i1JEy$&)Ob5n-0ZLu3NM5( z<*M%j>}B_@p_ZB_TQ`p7)VDL8xQ>|D5^|D4lyW0UG(1=u|5AXP3;q0doI`s+_L19} zn7Uw)(nS-ACh~lR7iQAVJ{3qsj#iR)(hg3d%Yqfw_MlNx#$xs!D!7tVDv!jX|tx&Zz-C`GAx5`G}(i6CAs`QhN_t^`hJiYFzQy$4Z5I zU)*FaGZn*CYNI^u2mnwn-k^1lEve=Y_of<5AkGytkGiv>IR-T#4lIs@t2!+h*Z>x@ z6OQ0vMu@rz4$tdT<}Q;4E{*WHwsrqC?-qoOAwS6RVNW7<_I30)N6$0H7SPPW_#+5<7&Cw=b`-V+j1@w;|V?zM|aRl(*EHM#@h zK-J4E3$2j4k5Y+SPMfN|Cn^d)={zdK=SFEWkRz}W(k*e6M5Ad&I(VLh&qSXriv1|J z2EY8uZBgBq*Zw2%6n{t|@KxWUj$r<;2j{oS$g?$k1Y?;d3fP$RHf$l^O4II^QmIRF0 zI%m>1^m0=P@5Bcjit+(boduT^pAQ33Ed%epv^5E+%6k2mJ{_+#6zEGcXgJng?`%%$WpLx^s&)kZ_WAqIG(*Xe1dL{@u7i#whhO(7euJl<1{zA-GcZ_p#u5u&=`OCiVfzH2qiLM0ErkdLbL&#*PzQ+iQ zl5X{5Dk+|}nD*sIP4s4<8|zMd{|o16(tM+U-I$ zLW5gZ6EkWxvH0MInO%IE$&_x2uom84V%>G+-x z+8Nh`{d$k~s#e)6zcS*EVq&b(0nMinnzVhjpFk*q310k-pXV$1#wR>HapY&yYF)bK z=68n3qwTQQ5nc2JiaE*!=+3-F5JWCFm&p1o6=BYNi-Xor3bI0<^JP-ocCmrU3BS1x z@j8%$O#dZ12tx{tI;8e18vYRgdJRZLP(d`XF**Ld;3_|@wA0p{~FSK$* z9^NUlh~z_Q2SwDMpSuh3#`TLpP&;xSN6$P<3!)%LAJExW!a1GNK@1xWFev3w%fAU; zsW15;oIe62+JsaStf^vX7g=X;Q^Ff;4Z%y^Zb5lfBQ|p%Gqf*PlO2xh!@N zP5ZbcF3mXcnMf$HTNT$j=X&ds6F0Yfm`Zi+&Ecu=5;JG1Vb;p8{c`@dZArB*!0%Sg zpCQEHRAL|DT~2AG6xtgi)>>4*=m@P+O=H$YnDT0;06hDXSy|2Y6Ka2M?7iCT^r#oK zdxHZ1B=pl6g=cY^sF}8XX8X^p!cJqLig2=W2{=rpzv23ZQFDtZbX)}i{EFDtp`D-F zJ~i9lzWEcbDqmlGfsp9mh%vhfyKZutfBCtY#T#%oGPI1SK6?mR%zVsG2dnfapG%Os zzo|HLKzZl~RWXX^qaKKRBuwTf_-{>9?1^U%|GwU~*kqmroIX373!B2Q);GP@ymQ3P zv?E%-#?ZhguuH-wJRmM(5pmA4QcBLE0_(oDc@89{l?yE_M2wK&=ZR$>?2#Bf3%ksd z15v3yQg~Z4d^vk+%BJIB2lma*@85Cj)|;8+ga9ON52n&~wL~qd4eKJ8jCV^BpT3&V zEc2hkc|1v#vqAhjCzXUt?dP>$cpS~96P_BH3(+Zn$O@^>N0rHCF0b|pmVzLL-`|7= zCmJW%qwFWMPyD>r=oi-Nx?ln7;~K^LRC4<9AF6aHWx#F7-oXJI^TVTW^Pw%w|`gyZJB5sIE9NW_9 zpZ|cS1t;lrnOs(94L!T z!paH~z1JsyTGWiavArR^+4xXX7X5;6K}^*8b8Nj@hm%4q-hPr zK3;m-i?O6Ed>OZu{bvVn+8HGU%V$W${rEhHPjqG|Rz0tCgduQ$GrgN6L0j|Co9yp( zYV*1y`K*r}DmA;2)OhnXPJrYBiPI}`BvjbhR2{EXC^Llv{L$)})_=3y%P9jPkPbqU zw?%Ts+357iI+yriiwu@G>98ttzFztMq4ni8M8|JT{wt)<{+T(x0Ml~JipV$SWD7O*Z|_=yNa@)voAt^od9WFFlw%GzjBpS+g4cgMYT@}2ux)96V! z*GtifQ1xWE8?oKX^!a?1JWJimPmiWwt6q(Z+w7IQ&JsV4s{=MZM!MIH?w9YfXeKi_ zI^}8i*6$&$+xa+@7L2_8NsNvy8c-^%|$ z!y!ut-=9n-z@L-8`wqmopSP_PI>QU0pxhsIIu4}zaO^kil{NM4C4kLtrzDMtDI*&r zoclV@oOyotA(eAW-v<~9fixY}b)F<1hab)xB$)VpTcW1eJ_-4)Q4z$l5?0_lfc4Z< zSfB`$g_%_K?q3s_XiMK&L(ZgNH!+zm6lOiptC@}T^vBwq7FFy-OP@L0C|^NC2CjaU zsCZ>5>nx-(gfOGp4zuQPvjzj}Xg8t&fVVsBC$2<{KdqGGNAv@=`?JAr3NoF0apUHo zD7ZuuW5;Z=pi063cKOS~{pI|T(JlwEN#YWeCQO=N@wG%J$YA^ZO+*R1Y1Q{2b+SH# z*5P~oxgRpO0RTuVkDj$LZV@X zGu5~Uuc{OvNO$8D=xE5vx(cwk#xKFwT)awu*VvS}a?4D{jhFQ++@wQF=|MnY-Bm|4 z3e&g);%&!f<|AA9ou6RlF>^v$QHh(2!ujRc;%e@V^rJ<3M9o*j}Em{@732D#(Jd08zy5YNg-80TO&M zEGY10yc4mt#o<268G`S`4;zo&m)_iBy`e+94-ehQK2)N2b_s?WyfC&!cr;EDrfHrZL zsCuW0aed`p3cO-zoTjg@k9B!6&O5#gs|wy2zLq@2D?>MZ<75W;2YewRGu44F zWB(4OJ`ecy*)JQ0H zTFVX^SYYnrj&}IjO7qz0OWv5w8^cNkqA>GhtXsq=@3UVM zqNaKEr>E9Nt7`>8pj*6F^raIs0Y)m|QZmf9vyQBa@YmckuzKDit6rBOWRqEpJXz5^ zdAnY<`}}>(tupI8RaUa<$$toh^R5H_*^M}9>PPe>A|1ee5}p65mvn0jDzA`1*ynJA_t`Qc>CcFnmW%fuY(_^erdX|d zfp#zM;+tx$ePkWI;EUu8N{awJV6W%haavD<&49PojYoQy|4mHK50>LY>ZKNER$!=B zL<&Wg|2p+g-iyb(6V9rz-g>=JD~8J05otK|8>%%12h0h>IQ$bNIX7h7k_l``Rvu=( z=^Bf>W+SLAmMLTl_ljrEE-^Eh9)wh~1tUJQgRR~g3;mC}O98(=!}mu$=ZsloveSC==H zX5SO$iBsuWBX5&J71Uc0k)RBX*BL>c30M}!L-C{Y7CQ<>QHHbP1( zjFzEK%L@;ygC`wa;pPWPP+}p=dO;_ zLezaTb%;LW2~`oJVC6G(N9PC0nOEb*mHk-NzdZvO6Zkr1<$bLL{6IyFV;!*>C75A9 z{4`Z}??tz^;sH*^h0&W@(6kl>K02GXE>Z&iIwyJ0(4};37Cgn37|iXqUw~)m@v=Z( zRC29&USdggvKydR1TQk>^VpClEqnF5W^WD`-prJvV-2H@eqVKB6+MC;nioHpfrN1& zuftwHF3TPGV@}rnPO2P#v7+T*kCnPZU=mGgIBy*8-9ZsZ+X*Gw4<$B3`ATe==Al9e z_060U1foQIEVVU@%22G_dO3_oqyMEx;SD;0UqWns@OhUiCsbW^4}Z~bkzEi4sgQK@ zy6kS=SBRvneLcT2=9_os)jH;VJvAE=(ybHqk6ZB3tD?szjhIkL^Xo>m&D_6#D4ici zgsJ%4GgASNRBcUkSZ%%fRY12&li#0dUb@UzVuISVJHXfTGFP}M2{yRbRyN^S!^DDt z!Y>VowMvqZIxOZ2FWp!9;> za#8RvSoi4{$ys&aXBGcttz1D1VG3lsQ0R%~u?VqN+&OZmQd%@@wT-<6orx{{)H{29 zNb>|fgNg@89L^cNhZ8fqUC+O|g7guO&FWMf*RzYxXDQ?d;19066`TfPEm81n0z*+z z{acf0gz!QiS;UN|;)5>a?lm~Z@%=`7wy4@k0NF)k4fNgQbqpN1)jrRU!ux3Y_ zX%WGr#&^Im(iyRZ|MNX6UFF$@XDaAU)pAL3t1+a=ashFX2+IGsZ+|3`zHkJ%;v~hT zmT)NOBIFkO@U*jyKIpX&vG-bjn6wW)N+Ik~id#eKAwqEHP=e>5Y{JA=B%`1&d{L;O z$~R#V>XvI)v<)qYis*tqfZh~0 zpV$|1oDl=)vVUq~Exj&;c%)=;hA3;W2sXhwSN2=?Ixmw26i6}#KP^lVlkW4J#%os< zIc#E}2*?R^ptO@%ISo-Iw<}4)?^CW0JW})|M{`02(;KMjtm*y=rcT=h&FniUOr&(d zzuKxulWTOQ$kS$j%5qD7JUGhd=hZ9{JYvK^Xb526Skgm(9&fxg?wBIaeB#r|g3lS5 zzry9C(<9GB*&oohZ6CZ8@f;UuY(zZ{BA~} zr1N+EF~tHSV|1S4+DmrSdsygRybn1=v3af-MerxB9J#k0sqDshWydRo!-|MQHzTEK zoH({}Yq$_Mj7czqBiHw79Jr1j5Lz0a5eT->1a)dh`W_Se`nCpOR-TV(%`ojCQxD79 z2VUlQY0_s)1+6iX8ULm#F^VzEV8ObsyV!&uKLglPZZ~$f_tN=Z;qYqdw!RLsCJ28u z-AOHp+~Y?&L*0;jOf57mR?uS8p%DY-q>VsaLbdN5$ z>7?%--cCX`k9Cuuy7O9jHvEU>!?i}M}oz9R$PTgsHuSktzAUcMuA z-Y6;51|AAo)J{O@H;K_}TMkTyt0hLEbk!Qg@@{Ri$szoZ2f(*s;}?f3OY~M+hUv4dt>ejYH~Wa23JrKD4C5|d}?R^brRL?5H?tL zpC~A-i&YT6L!!^OVGnvpJa0Rk>WU&4FeD|L0caMsHTexZ1Zg?U*L*PIZ-bC>>){RA zg-aSmuGA)IRcrGTlun2bw|=Qj0O4o0zWl|KSJeUB&aIz!Gqvp~;La5KrxJJ-&&u%p zkCZph*|XG3jXn?Usy8O#u(rej`g#R95%vI$N=4c{ai6zlCxpp{I;L=t%{wKP!g~7m zsZ@tore8z+e9^uPeK|^%-~`bO1ypPAjb`g%ui4EGr~hm>!$jN|=v2`$AUi{^(|8s- ze}Q*(pE*>O3=axHMOPqE!+6APU5;j|$fQ2aE$wU$t*-mFnj1*=AXkOq{I9jidQT@~ z7RUrLzDDuhy4q=@wo^OOXwXh*Z<)-m=Sr^@Py@r|t)WzRYb*@9JvnCE%Uhe68d<*p ze+Z{11RN6UF_87FF+BWYDFS%L+1lJilclQCH&4}LbBe2jchn2ZC=F4(W8}?EPG4;E z^PkHvhL4nC9RFb9NcvzaTyPdWMEAB=V-0IDW8?bi9tEJ^%%==a3* zh#!sN@oS>aU`v4^1`r=db=9cX5F;lD38!g-e?UdXdlptzEf3iW=QW!6FPH__%qoNQ zsoIqKX#76OL^`D{*N5iJ{xmT#+Vbe$y^^L%oW1;`Aw8`A9@AnJ%md7P$@5Ovu*M`t z=A~fXhxT$auZNfMyk(D*<_xnU=K)noci#FH+PRYMB@ML^9Rgi9-~Ors8X01%*4lx# zHivet`^QyGA4(4%IO|p3(u~miua`*wrRP|&?&rZ;8%d?7Y!~rAsm;u^LV=#xgXZ4n ztXP>jSSOYOwjLH%d#%=!&BZ3vAr>nqL(=h;Cufa zPqpVo`wOg0uOBCJ`pp*RRQ`YkcruyFq6&*S%E}X<(@%AdgUY^=TsAhUCGew~Wx>w) zg${)8okCEm6;%+Y)wVWsmwT`f`d zDvXEcoP8dA7e?#!bt@$UC68si^qDaSld`uK|IR1oDhg}a#o+YC(Ntv=jgR9*Zb_R( zyEZ*{syWYU%&<9&D9r}8eANzJ$e$yFdbH+O9Nrx=L|9noR%O;L*zN<$T-ux$O#L_?G4pzHZ7AuQ=&9a^sk-=1?N4U|Okg6zR5@!&Nbc7I#@@D05Bd59WO}3Rl)d@(1)j}su9(cO#TB^XaRQ_gWqtW0aajZ2{ACV_XlzJ*;gBw!Nkx(xo6lf zccp7NRh1&Tiy4~28QL1a0hX#l+dmY|$A)H%cOND-;Ie2*gZ%pk+D3j|!=LB2A#8GpcOIfYK83W9(s;f> zm=pE4vt`;Am8GHXrFGF9LF5WhS?LRH&~|W>aEGiHkjVTlq$ zt@TVf(%W&#I9>J~%rLT4D}`Urw2)QM1AmZ&`+vxVQLnZ?v(DRfAD~VdT?*op*uCD^ zHV>p1;4?COp6r1XlRai~=**VZYh-iTtG=`E0bJTCTP9s=Z5&pL+R7gqt zte+)a7j7O4&A}K!^VH(fak4iCD{M&>)yu$~slc-N$yB;LH zpNl^;m3zD-8ME{g_|LA1`8I-P>q!`)^6f%cbTpN58$S6+BHvkkc;AcHksj~VrgBYk zV#$Q3(Kj{?PXBSF_QZZOIi66z|7`-x)tVrIirbTlxOnjC4I{@Js)JBCeOb_ilw=YI zOPvl&b;583C+BXw{3PI@lcC__$$X8>c~gt3VXtlvt6_i^ZZcLuca32U&G}1{WC^l= z%0i#+Bux{BM6KSE9?o=zx2w8XxrW`kyuXd7(AKUuqg(XS|2#Acfo9rEHpI6pw74PM zw5~k*i7llQawrQf`tgDpiU(M$uMT*_OZ{A|DN%qGk`LEvwAw9R+fjo!ZS>t^xUJfp zg=_*=e*P8O^m9(#0n!Wx{okf82S&(ig%6fGcZ;i9$C+QkbU*WXaBx%wtadvZ02a*g z3ZV9(dKLwAB@v0j7g&8-#Ve1)50qwKk+aq;pVO=UfCt^_7{~^8Oar3zaA8beWISrM zizW-i(0K8w|n8~(1_RxhS8cbCFgU5xVgnW+I=%%ae9lUC|jMF!0G zWMoh5mr;Lkd^0+x75a8ICAn)}?nv(InQ841@iJuhv91D{QikIH84Ke1pZWHs_Vz?! zj|mB$SC*MQXAbVF7ayiui(*{G@ugmFjRww70#C!3WRvZEtZ|%Y+LF&zmJ&XB9!E<< zhDYr8_xRC9-9$JHJNbQBtPX%xtXIWFMQM>!w0DHFuWP?>@)n5B$;R(!i?scG|Ffq% zz4nIN_B?1Y$@ZkhJM@)ryx&A=hT+$0v&)IZE$-4UelAs)i878(ALFGvDF-emf#%1*1>igO~IVLM3PJMEt+LSK#3_x z#noOl3Mm&S&3X$ElTy&y%SIvc| zcER*M=->&#G17Wq!2nUf7$fY+_nGDp$=fXu$@~kspH39Ozkj7gB4k;$C&!FV+iThJNZ9uK7+cK+(#S#Si3w+RYUd$lc4%|SZ9kV zZkUgNJ2j()i!}>< zyGj2vNkV+DIX2WP{e0j1t}nWM_xYZS+olUzUHqZxW9)}YaM$r9djUTQ3*aT_2Zp@} zBpE}rpc|Xp63M5SDcO%(VPnfk3-~K)*aet2h$rf(PnbN&TN}t ze&~B1qkl3Nb3ZM5wZYqm&zmXq_7tavGPo2&fE@VFEOwCuhW4oU_Tpo8n0%!{V?mkN z_hctA{}h3P_C{-14nBYqG?_eKc-za^k=Ta-vqRIxWSe{)hmZJWSj&dA_BeA!lryr_ zaR+QZ2;Fs8r=PN|4U8^P2SQbH)Cy#u$>c}O*o?S&9vX)=g(j{ozbZ^z=cqG&%F(Qn zMSSNpkJZNQ5mrx+ez8=``pTB>hmJcik#dZ{qN#cA^Kt#-HY&Lu;}0k zj&(NhAJJ#VREs}Pr^6G4nOAwFktHr997mUJk`N(K%;qn}n5*+N?YKj>el)U#26Vsl z==rGIe{oC9Fb8rqXo8Y0Ims=%!cLa5-;zB-4~wC@KBbwcPli3?5My zR(H1$fA6H+oo*2lsn5P8g@y(k)y`vK#Futi({FFSMEzH_z)PCsZVRV;;k3;MlsOx= zrxa7P0n#qijHTSju`ArnDh1qS6aZ{VUE?v@`&2^&N-{-C8233}Oxdb9x^6{rGmZSJ zdF3Q-Ss@@;6`l6tAdU+QGB9N!^C(mr&D4~2Y-@oM@$Yhoo^xOd88DB*SYX;?B3+6d z7G<}L!E8i!rf~p;zT-B6EOm2ZgphBtHDzER$4k_|RueU>gxnma{~@ECd3$>37M68N~z`2dEt{e>%yAd&gh#) z3z9ZyKl84Zoth;%oc1*($0g@ypxN0Y6~MFQFT&_sTdt%ytRVyYazi_pxP4H84{gRF z5^RtmhDcsogz#LfGa<0-H*I?ZM)M-|{;nF~KQteU(!nI)%Aqe_V~+#@TJajt!+s}T zk$6VNb|}SbNDWutsC{JV`7hB?!b=_($6ZPX9=t~Zvv?-IytDar=OKDJ(Vh^~4-bP& zN;2Rlq=JG>Q#eaHJ1t6|B4tEbJP}}K=DALtQ50Kx0PUvhYh;UWpg=L5roJka@2rbQ zoYqVKi&jKIVC6F8%I{$RbImxJDv1gJe>=JL7TD3lSkA4gthrPySukuLCMmyYU9cd9 zXbbMWL^~39TONhI^~kHGc7%Cn&0}YM3Vpm2lV7xGjPbbKiD;$IvV>5`++Q{+JmG+( z>VIv(pT1P`0+V}L|5Cy>_;SHhbeZ}UWdY~WJE;=zBY_~#(P#s%(*E`WtjPPvPwsuO zAdkq?(%q+$8epMLp77G<4q1tbNC-)Y3+J2T@pUiMz}R~ZLXj* zpi4k$9&O7j_uSjRpiy<5uGFPUog&Rkf1>>tzNF$}3=Y41`c%XPH+6m+Mfu`espspK zIA zg!tIIUgu#m+MYXyYtM3gJY#B*&Cg3Z5v>4xZ;H))K% znR^;+ZpZc4=IqJOvh8|Dd7(8ct+*Ne!gv3nk2})HB6Q$1=c0Kdtenm+;!V%P_h{i) zO<0EOKQrVgzrL>2XV-#`QF(2Eij{z2} zDZx6q-)ziP^A^-A@*z@@=w4d*;-9_!jJUZr(!Jxx(f+E7qyD1s9Oab@mj)`i%5Sa9+Rxj_njje+?`1uOnjR#f+)DCy-9;Z3FANQhT1 zr0AB)9{WV@IyF;@t64~7$eC{*%V5*VrqO9#5pLfx$vQ|UO|tFRzfo~4L7~P!{|pxi zzDqr?mAQCl`LiXFzsp!GQ>EG@>U$YZW`97Z(5A=mljTY0-RI4J7j3C(l;nqR;>l0@ zg_4U@Lbjh6UG)Sy`?LzB$Nf%QkaOT+#Vn(3o5I^eh2|%6v65tpQZEN7b))-!GhtOi zYJk&Nibk%5>&oe1)0ZTfE&!6Sz=yWpQ&wwpNO0$s!#B^b3zX`5z0SRscyJy0Ii-_p zg`w-%Cb<;cmVuM;0g!7{>ZDv<*%(Ce_3Mare@PO3wR?0*Pj-2noyj&BxcywXSo~n@ zFOrgeamWK#+_=w2y8XqjkLj_K`#HfE=@&$p>y@yLs`I%8i{=Bm`wnzcml1N8 z9D}BREXp9dJTm^N#V;Ey?bkb#?L-L)qd_{X#c8qNp`w~y&0X?NVz?6?(IAQ0MQcUX z!_kTXEoy*cy+(=nMaq5I(^K0c%;nHOzEk*(LmS%od@N{J>PzZERw|6L+jN&38|#Dt zF^qTz>+Bus5ViFx>uUccpGKtOnl3!OdRWF-%D+2F(5BNkx=Sj2gOp+QZTTNeZkju^ z`0cMD$^z+ZItYS|H>NPsVld=mTZLEH?gJpjG*va>?mmU zsF=0hCHCIB994RMZiWNY$J!wHtv))k@D`b!zaneYjs;r`!8bxRO_*O=YMvjOQOJQKM>YtOgS;5K}LHeJQrKp;!_484%xm#W5{#m8Ft zL2SK@$(U%1-Nhm++?Y`{oYReocdu`aTq=V42{SoxXdyXUW6>%7SPLTBE;Bmal>(>a zjuN7g`hZn+?k0v>?z;B9Z^?zSTv%oysaexh|H256>+5zFfQ*-E;SoVu6>K?$G`Jv^ zm1~!6LJFKO@tm>l6g z4LF(Nyzgoo^8zv=wL&%fu(8mz21zAdkFho0Xom`xm+$IzdXYxLP4-KJZUQE<>^F%g zP754in!QgEo&{gp?!LI5WtrsgS75JOYh3CsNNO0rjSme6NRFfYw)IX}yIQA=upq=7 z%s2X`f?}oD!XcGCfm`am*DL*R$5tZS(t*@8LrQa`b;R%CV9WWPCB}Ms!T+rb=bfc0IZO z;O_d9-hgV57S2lq_jHv})1IxL0CrBVesW_ap0|9^FB5_|UwyMKE zb*R{;Ds@Y|nZG2T3(4C?l;?kf{)Us2Q?x}!qWLTrwZm&ZArgqi3>vtg@)F68`Whxj zz8bZ{mU8d?j>italdI6id{iM(Z2hM6@c{`D^Fp*=gQQ0u{QaNBB(d(L!j2BKYiMX= zeCun3a-kc-!>qi~0Y@ADA=IUHb+rv$ir%#RnO2-m?{kiUW((gOL-yL&?`exMAL3LS?m4Uza z=O^+zHyV7~GBrV)wk#Q5cQ?y|j}@1cR2v9#_--9;nAc(PAmeg%G8`-pm26PuhC*#q z$Esa(6l8zlU?lk_-N$Y|BFge35J>~od7=en?-;_${1r2HZJs|Bmy{E_7QqAIW0PRX zzc)mmX24ztL6OP}oUE)5L zbCvvIbJELOOi0aUY0)E5q{X2c%f3ca`s?*JTdHIyMQ(yuk?X61Om@@LkKe$$j+<%T zn3j%x0IaTP%im$~B4ODh$V|awS_#jl3 zJ(627+;&{`V+IC4p8#HOLLbQf#r78~X3>-Usk#f<^2v^HAyEbvh(O6DBsYW0x;EOC z&qC-8l*3m{G2l30tCdpg^Jf%90B!;AXlxikM9?uZ**6wMg%0GHnAi3yP*vF${yg8& z_C*tf0mn&n9KQL-hzKYjjUjrBEKwxcVT9;hGmD4OD$U$3N$*Ace8;LgV$y z3zWFD0&%`-RQcxt&bR&jqtgdd9(9LIV!<%I7t{7LddA=9whUPB7-Lc}PgKR>r@+a` zI++5c=IKw|z~&wFNQL(Xe-{f8%!CH2IKIuR4H>xMBg|{WB^89Y2xaAXf&5C6t(~%>^?_T7&zAhw_pMjWktWSlhql7l znABm|!sXdEC}F2>cR^Rv5WRS7+LcJ*7KxrB9o*Y&=H6c(`I(YJR5fMf%u^hw02zub z1=BJbNNGl2MVy&56$j6WzQN7P^cqA zi&yN2*7(t_naF-HW+tTGpKEoMN2jmN@A%kv`OXBo62v$Ua^-lyoz=x)d_!&{K1xwc zAvi_2qs+8+4Y1Wc#Be8$r1Pof9j$KfjOY_4j4|CZ9~O2>c94p3Fuv)u_+^XURjuS zxI??L!ySb-ZGp9~qZDyLL&kGhm|+W@xYugjSRvZqNFVPJQ|6rZ#y_;m|Az&jCQNJX z`T@}B@}ux(85JRt@j0cCmTWo5N)r42+5)Pd24w60a{zu{RzU1Pp>~$mD@iJj9Wqtl zWvZjd7{(0sF-@;$?k}89del$rN#BUHtlpokOxi5p>qd{z z%rjn@_z)QF5L!mH{UU)zJ@H%?e_@)NWo7^VYSWJL9s^@b=KbeNzRT}tF#Mw(zn>{U z$#j+5VNj#iFQc5<>lwT!js;x67W?S$K~13o{0*z0I!9gvO*)0;mtkt!!*q$cI-Ha*xm1u zxA0N`FN_934_q|Ct10wj=_quq$y6&a8+>zE2^B`={0CZjJiai@_RFxVo!88wzcE66 ztP)Cstbk1yhLz#0GaQ$My{1Oa$@BjcH`l83{Ko7Twg-h7Z9B1`L^=96tzkuU;pxV=r zuG+Q7C~#-SpjBvcn}9jPUw)nEm0cgfCki(gsMeD2VgY1q%y)7L^q47C^Y)P~Was+>(C9RkEPifw0oDpZJ{kB5Kp6j^$rl5nCld2>gT11Tsy;2zrW6|W)s zwYdc|H-~V7mr4RehHAJ_Tkj*;hiu8ameRNX)X(b(5m$SE2oll=Qt~t5Q-Z_*5sE|fi1LrMEyjV4WAq}yH-0moM)dtig8i~7OL1X6}`a$N9nZ|=4 z+Zp7+fy~JZoh#e+{w*QM z4GOOCiV&A<7vSOwPXgAlF>Z@k_-7@8;lfr8c|{0GH89eJwsyN<$@x3kvQ_a>CmzFM z>`{s2LaFasWhx*5w$-^9duh4Lc$u=?uSHt!k9z^Wk2uzq%ts*on4zwH(fjkc9)jMC z{@t8Z=GIr9wjCG0vv3iPefOBRn$ml))FvffxSjX zFreG=N;yBwx#NBvrRsAcpoL~%3VP9zY4lD0`ETY)Dj?0N5eJrt5AcWKmu+^R3RFkr ziQ;p^mfpPyP~m8il}ac_#P(v-?|Xg2K{LjaPoUC<=N?%V>i6Kn2LHR-M}-2mYtxhE z1EA0?C3(fGg3D%6+kpNLfLu5ctijvut9pZaZspJ`>sKeK0t$1+l4ZMFj)yqvEXl2+ zqz{%LYNwnEupVcLmR+SRf;t~sQ3y^t@faSS1j8Q5OnQbMZto?)@W)rWH0Qge`U`&? zVsmsw(8(8BYQL|RkNFf&{2t2mh~`7m?ydd)EUf7YtrZEJ;6GDr*@_SSJ?^&PhV{HO z0I&^^k-7G6f6~Y-i7a z18BEj=8E>f66fOc>g(X^5K38}u8-``I$tc+xbZ+Isy{LZi_fjK9ilFc@&TA#d2bu! zt5{2R>g8-U7RcewbiICjlQo6fxJUA%jrn5oT_e$awCJGUH3DBxG(nH8C6N02;=zsP zfjZ(F`-{ z_b8$hkvBMGG(J__5G|VI`>@)TBU-5BekzFC$5b%5OA6EId z`P);)<~L8ihgO7LS1um3pjb=`g+1dP%3pX;J6_#k)1ANsCaluT3yB+ zhG8lQO}E&|x-Zg7B7tPvFe`%|HIv)zX$V5Rk5VtlL*BDN=D=o9@sHx~%+ylI zXdw}xJ=84hIK^4yYx9#@M+tD+X3+v4M_*fz9mhG*3#S%U~Q$Oq3XGTZma12;F`1qU9vh64#0#T>XojM>ByeL!va@yh5xCoQq4kNE`$K2}+6pOtE^BQE#dlI(UXb>ur z4(m;Owe_^4yfh#f8bKNpMxdbfhQCKsd#z^eJq%qkGzb~Ju{*iPw6u2T!=oGQ0SbkiPR&DU}+Buf|HW-H5Itp8aY> ziEV;x782c0l~OSr8z>5B=z_4f`A5maXG8BED*nUf166~_oCXGp>kFiO%AZuU7pq3< zT|%X<2UUqv`+F>jfdH+1cZYt#nm-kElav$`Z573V%S}t|HN0GWFF1OA2v0WJAgMQ- z4>CUUS`VbhHJP?DqalSG)UFQ$UlESQVYP73D>85UkI%3mh|!gp*a8eB%J=aQ8C8C= z9`fWS6F?Mt>sPLq;4`nfxwal9}|3 zy_j8|N0KZV?T|5Jag4x4pL}9n|h-77i9bcFV93j_<=)WAYTMmam zukBprdp#a;@|aB0mI~2@>*~Xp;1U)MvMRj?CSii_F&3Z-Mg0>$)Aqrn?49AR*?)-x zi*yMoY$n~k685($4mXN33BRRwveJ0%EMn-BVLp1{?u9}=IX`CRjTlu5)_(q6Sn~IV z&+_SMXcFDBnj{!s4h`|ovF$h*J4F$lQ1uX;s*wvOz?hQUIbW3}-E=^g zJ0C|~4Pifd>t@xNYW!Vd9dWy;l9WmkHS9JsF9KUzAaV#>{q#RK6%pZxSDu7yn~>uY z94N*p;`t&cpEyZ3;oWm3iv7ET(o}h(l%dez3&IygLF)tIARaWduq^Db+U%*yl`1oI ztvBB{E&b-Qfjk(>3cnr3ZwMskicoh3ll|*X3ekP)k#DV3+79wd4*~Y+IU~z6_pUvP zg|=IZbHRt=?Y#I;CYw7?=sxc@Mg`(;0mZ_FIRnAqSm{NFNHRH!{~h4?fK1VN~%@4 zm94k)8WO6hS?}(Gl0Y#UFWGTu1%kJB%nO4)w3`EQ+e<<;AKwbzFGxX=#kA0WT}U-* z!wGw#j5~7y4>a;^!22$wd8KCWoakVInr}WCe9Ml>*t6drx-QnU&sj|1jYJMWIxX^!uxF2REu3CK@2c!(7(uz#H%?kI~C)j%Rg|j670a+z&V!pLWL|$h`06;w< z_m3xa?KQgwyM4qP{#;lgkiC}j@82NzBk~Kix0!<*=t-p=(jm|~uzUs*k zCI5UQVx;`+BJg1X@KRnTi!lhooUG(}spaRxl!Cs0V0q{)vpb|s#aU5R%1UFkZCwtE z>FsE}UOj7GB(~wyLOdg@8iV#T4Qd}v@HO;aXeC36X*=;{MJYS4_B_% zUHnxNe5B0%n?i+wz3;xZnVjHvnUgWubIzFmOd;aA_+cXO;LLyAQtSR8$SJJWGm%oj zOJY4Fb3DPKGiXF&>e795S3UAJ1Bp)ftFi~0kFF>Vo9|@*t{l7EJ3lq7Ayj>v*}b8E zqF{8{gAD%O#ly#6Z6q$nO){$jZGh0^U!0#J!%YF%{e#plOJIrN48SH|0Ky(1l zzN`y5mZ7%g4YFr!v6;9`0AEu4pS)qaJJ3uW#HGlFR7bVJ5$p4Q<7E!?$HCKKKbju**eZjyb2c% zbUWpR7Pu*|VRG?XOvVmbM4l;I#0Bxcow1&s8gDLg(uD9I2ViW5kZz1SANz0dWJD4h zA&KP7b<0@a+~wL%-t{}D2h4viloyf|^o>P0t<~Y%Q80GlK9(47g2+s<;rHstt_qRy zx5Sb@`+w(xJJ+2DQ^kroclyzxw@k^hu$SeRj1?Cse*q=XtqWJT7fhQ1F*lC8X*6+{ zwzmu}K73O|sGKQ-dJs!U4*srAX+*FJCxv_KUhzE9&CtgqbYj_eY;dH0oGleR6gPZQ z*I6Fa+!E}2wVm&Qb;4#4jQvcVY_l_Y#8?Vwn!dR|(eaQoVhGrL-G@L5d43Z4rO*$n zGeLJka4yFn@X`mrXXuPE4LsZE2l#$X;JYgN{m(4U=JlW|ZK5^TQDj_WqyJi?*HGUi zJ%|`NfpQ3C1VXp;XPwYvwwg!jgP_)?q;u^zP)?{v2(%h}vJ18(2m!z&j6HP~C2E;V z()EOTG6i{+D;v|1%y$hP3e03g$Ra8nv3z6!h-5riG)m0X31h$7k2Fif zv(mK@1a7%ZY(^(U(E&wcP?=Q`qc%<8WG*~rybZD9f#wgBFyDPooS;}ud-myOyjfzp zk4V8I`{@#Q@kyE_yxSq7fLY)(WOtY~#gj2qcfdWQ#h_Z&bFy)6hV~Sq zIrpMTtk8YQ5?y2CL(_}axM&5r#`Zs1AG$uqoII@h++}?7tbmKJI8okuop(K*`>FcM z_r9bcZ!__6XjNqh>!*`t9X8{8%}?~kt9HWO3z)p{)4)Kc6mPesr8>Cl0H3rW#m`*( zVtYdR;Nybz2GKd}dZO7}$Dl&n@98Kk$NU3L4U#(S!T8bul)Ddy9ROU7=7S%@^EN5e ztkxQp4W5`5YxR^CN`r}2*XBu6Cg%Bn#^6)7uqRTgOHD1j&mKv5z&~gOJYVt^s!<)t zGe}N(1=%>>UVe=tx{_H`Zhu`I@@;M3vc-cvey!v|``k85A+kwJG(f%mV5ujx8~)T@ zX;7Z0>$ZV@=a)<)E{xf~C}g`9A&q&Jr)?~Y=*6!s08)CqZ*7P+8e1(yw9jO(eM@@k z%LDmN%C3wzEx7LuCezA9pQp;*aD0~2J9qRC*NNG^9-{vFEUVlSW084{hPHkm(=-PB7ch&K(rr8=3A-63}W z0W1lX`1m7FvhOu?-rEe;n#JJ=+ag$DH&WS={dg$2V!!l5kJZQmO(WDh@5NQq!iL}@ zBWT-e-K0?)f40_SDElSj$4)O^cDoU>#)nlrG*+xwS5TZUIktU5j3Ijw zmnGxv-4^Xa=~`gn^7#zXh;aTS*Pq$mlJMc!H!6qoG0x|8w9*Ja$IwUu{iG_n%SR%k9gT z;O#Wqt>&y0;E$ z>kHp~D^Rq!Yg@cf+${wP6sNeB0>z3u6bVq=-L1I0TX2eo;;zLlc!0>scYbs3+?ji4 z&fn+%J;}<>-fO+6B{h0OPHW8l_}_eQ9APJ1+e!|bHa!w zvH>e;xKjpSWWw)p`Jn%nl~^(RA&hU*l&uMc?+jbhR0CL?>DKCK@AbqJA8QakK2 z^4rHZ2ovixbbyyu7CqFPEi1+heH5z?z%OQ@^XdO!)HDra>W28LRpYA^r_=nKl02cQ zwqIXAA=0z^El^yzz@7^@CGmq#F0NNmHpVlAZ!|XH-8wVKMc46I!FSO%P~BnK{vg`* zkzbN7*9RY*fbdm|OrJZwIkxVNnrQ2`b&r^}@oS%qZ)J(9fxdiRbvZ=lux%jh4@a^z z#BlNwCAum17fGSL$UVEPdisJOP!)Y(Y_&V$B7)I$;OBh9LV^0*ySge*WBF zJ22pPbU4edT(64B;%{-Ub6cL-XDF#!D0#nAOM2}2%l>4wi-tjJJ`Vwv*Dcg>H&pVp zZ_irQzY#vEYFx|pIX%ieCLu=8^Sx4N^%+lXx$6Rai(d$~ICsL*F*y6Xvd$%i&$7-9 zU%GxcP?;WJD;i!E5~L|r&8lz~UDYOcG_yPrNs;RF{To6)8upoK^=(wpWp*Rjom9K_ zqVr-!)cyMb!RhGcSn}15RI>QuXYOX1C|Y}6 zo?ZWmzawqrUw!*j6MlOg%@jwGOvbfG5VM5qOd1ha%Qz!m9vpZf)f?qx{IRN<(3D!B z2YyI4Jm1Yql_(%e`xAWjBq|sne(C1vxs84hE=a{sye6H_>SY-W9nqXkjQN33-TUA~ z&iX@A%B`06lKHz9Azyc~ONQswo3^sKL5qf<)UYc5h-dZb7Fmj)NJ~_wyumKY?q_>71*RAUss3J+6YO_j=g z_wiWenl{l&;}bpL^=r2ZZqez~3fVs44+Cl36>wdj9VGfS8I>=41^voiZ5xEm#(&Aw zt0dfOHORqr8?^kSlJeZE2qZPytE;WaFVxG@vCT|8WhvQ5=1ib|t9<`~D9(5)cC!0Q=Ep@5gYF|NOsB9jYg2&VyZ|$x= z$DGiokbG|W(NlwLrg>5*-9N%txPn=U*Ngt5w2!vAY5XbJd<8=lDi60-u98SE^eKHv zO8=U&rF(kEGGQ|Z^t-xxiaN6um@r-wH0`$kYjX*Q-~p)$u(DS{=4)QaVWU<+|> z&cu`Aa)9O<#aHiJ($>}U59@338#eNPJ7d)K zn6^96S_EuAdqaw`#+cP+Vh8(Q>;^`XvbVwSH^GvO<@XDQjyAe?imP9I$3F;l@~Hx3 z`Ui$uG=WtRy6IXFnKRtbC`gq>id=rV&Cl@gG_t8P) zjgyAOTHD|AuWSv`7}}*D&0TM`b*ayFzhx@)PrsGvA)P!HjUBx|SyYq0#tKEM@|s_d zvDdbt>=~KQvKSzkn;3pK5w^c+=gEz|ezS21x5R=Xs2#idv4B_9KQlBAp0+nv|N4`>wO~}Y zM*+Q7-4~6Gl-T~MlWA8N{e@l-vk`Io@Eyo!2otEFswF%{Ip74A6Ok3FVvH|9sc!eO zGz^qoF~w`U&ke-9Awg)}h7S5uUSSU*BDu0e0f5_on)_=o?2I(%7i?vmg6&nu?hiv_ zeH!@^_KI~S8oX|a1?Z2iFRWWgKBB-mVPa>=vv#Tg=+>`uF}501bAcGkyEU!g?8}#r zeW|%bmPmi%5$1+qs#g^)2<}{O?j&5e?3{!HkB-X!5P+>)$)9dnQwu zV_IEJ6chM7*7VqdestzawH&S=np4@2J?myjp$~GG@V7Z)%w_2-O!}D3YWaDApKUzXP`ZbvJCx z^}o}v&ZBQHpDm&A-IspyTUlP&xctZ_jd?t4&dZ>q8X5Vk_;lw480lo`Co{UDSt_=4 zX`~rh;BmH^Q3peX3N{cMCKZtsNb(;tp*u#AO&T*{ojSH{zjBTB=`%s>#7ge#J=eFu zjujdjLHW#sNHBcby6@~hFdaR+hlg()a3R!Km@PAcpkw2;PS^Wj*O!m?LLpGtgR6a2 ztF4~3IkK800#ycwsp6fLwZnrl_A*fl`uk|*_}-U>N^ddGxE%$I##) z88*K*#+z$^tRaB0j{Z>e%p_K6zi)+n{dm;xnLg-ogKy7P7Z*%D_FGm8Q`kAItzpfL zBvvG!2t~{W@83@vG;2hT#5GUN@kwE0W91##)%X`YD6bY5{gGXsT4}R7kyr=u-|Mxe zaT@HLFC0lGi$K?wmAtD4FitQ92;*OzXuss>ecqqWxW?`>;vYe_J^P=Oux)nl-**?Z z2;m%TuN^idP2Y17{)TwkhbG4R)&V3IF(oLV2?3>t;wMQ;2D|OPC9%y#a&T<}F60aR zvOH=6C(MR2_n~VJ4G(c05R;PnJiN`734J6m0^+%1Lr%Ov{>O3ZaVsuky5(%;P`@R~ z54bl2G~WfK2+lUomhXD-GDMj5#cgGd0LUdToX&2$%#w_?A5?$$SmA0BQv+ekut863 zZqBx?pl1v(KtCsb6)V0hjxvdJ?Hyie(ZbyFetMd5FK=CPjyk@=nU)am1>-oyXBimB}uxpjwKRKVfo?XQwX^GsQ~wo+LrJFPWfR@gG)Dpyu|9`O0<_*lcym zPx1jp*0jYdY^^Z_3nccYU1g`6vi`v4lveQIFij#K!eJPc=M5^aW{j>a&v^=dSZ;dWsI!2ug|9{I$o%Er zgfA9&7Ipw~R!RrOrgok=yl_nZCKHP5+M&~WIO>=sv0@s>um+apu^cst#w zlB(C(|G`2dAf{RKg^ppNOG5}zN;ht!um?M@7ERHTbOAWoh4Hs7;jWjgZZ2x z3oK9LpPYgDjoeEUi}6I^Ph@$a{jq(arqtP)xoejF@%Ws!58K#}jP9dfV(xS0L39pE zIXywjm@NObhN(r_2F_iSRR+}yTPM{>rg!OlvTN|6}KlchqN5eFJc+D zVS<XlFT;x-(*PrVHaN5i@ zPK<)AX1r6PZ4ET2WeLl`#JtNS3^i@$FI5*sr zb*HT#G$79Xw#9~wvj(JFABdz^*$Ctc2b$@A-}q@Os*~tU$GV%A?f2>IFnLtq(Tntx zu&@$Q3#?6PS85DE18+T3ynYcAALMxEs=n&BW2wuR#j1XKt$nwn#F=7d%4O)YjkL1x z{~od6t=bj^ge3xIoUhsIlR(<(BnDs{T0#UcE}#N z$;uJs%j8DHu4fprYkJZy8ZEexy%^f{Z%4^4`f!0er^|MuE2C2p`v_&XSt1c))v~na zld_MoM^gcmsN`7M>Iku6WaJqt)wgsmJeD-5#4XGE$`4A#T+p{esU;VV)qu#)3no

(UZN<57iK5YNCbTw0uJC`nSRY`K39KFzdmxyVg)B1 z51S!#F8nFE)c7?`4*A(tphppGGTu1GjMsReK$Eq}2SYgxs4l%gtgNhaw~krJ9T?xj zqS`-7$i9g$GF1DNYX}ITwCSmmUDNx#GHZPPIQ~T6MYahn?|s6xL7=^2^M-mqBHa)J z8qb~&y+&w9gRWt^H!wcM<=__f{Wn0tG@1WreMs36N4}P$ik~lD7l-ldU)|s5b!BQN%>PR^56n=8ltO zo9}~5CBRU+9BY^5^Jra~*nXS2vE$iyODnbt-dI<#gy08W z%5Mh_Y~8!9JYRY9Oh&)a2_fAqq{dwB8~sq|*;JUNNps5V!1|V4%u56df(WBeU(TT9 zV@if0p42^iuwL?HT(=!ACys0eNa8GvjD1YFMKUlO;yk!36q z%VDw#ZaPN?a>IwVk%Czk$&?DxkkQHu%>9<_y|bXm&<94Y2qTs7|D@O;Ae zs!0u2PnIgiY*;3gNDU2*KALs6pF6Y}YMGoqaJS?*vqu5l(pM{g{%I@447iKFHA863 zQN;F5&drwnwcI((MDhH~%N?7OI;&K#l^fTucbUH3n|FSMf`x{_!Z|eQW$Oz}?(hCx z=o&eQ6@ebF*ki_5L*LUYoDPQ}doi{TYKRm$i6v(R@Rw*+f!kKe?~>d!th=DtG?c*L zh3TG}WI&wwUHDmpRWutueAyJaLY)b41X=jYIc3_2b<0x2vgVRr;J9kIG7%VaY;r?* z`mY`}&~6gqGyF3jKKu=J&XM#Br6~v4D878;{sP}l2CHLDm8u~7j?K+*K(AOumP9~# z9tB8Y$C|5sA4$S!i}5Z9HH?R4Uaj+5aJF7_Ao( zfFEaWYR&svg>nS%&hxRyb~nI@3dVEk+Pcr?>s*I6(0nq-yRr9Oo0jcVokE%90|TPT z%9lf&9II736GL+lz!%JSO>CiN1Uw#Nq%M`eTdzgpxW5lg;y6WhZ)aKI?wL5Q-YHFZ zYC587FI8K4Ft!6V`dc#=lEW-4A%%hTS>{?Z+r$f;!?`>UU+xR#;;y8bZo`Q+Ek}I+ z2pI^GMIK0UBR{Q@ku8-A4X2Adh6LXjU-TGSLhcqG)paaij8I8tXa8VCXX(m&f||ta7o{EdkbygkE>n&MQoBGt=PqFJ0+#Ye%T!FG3 zm$Ug2Ct&1N&x!VYI^#j&AGwuP7Nff1>g*hL*6{CeDQ6xT@nNqj>Z1PE(GjX z{RI7poHygDpQ)K8tE$+qn>tWajlBF^^TrtJCHHF-_G@#|HwQkt8& ztZSB9i)eK$a4<~R6E^F8(lHh9?%IQ#9v2v35%Pp7L~S1o(i6K{DY7Z`YLVSU_F1#g z5f`!<+m-{-mc#2+O*PwpoTzd(K3~p1?=Y!{HM9ciM2O&HTQYqFmd?x|MUyPWYaV1elBT&K^lwW zh_KtYtqi`w!_llpfBiU8z5i5gJU|0Ks#YmZ>Ma7%4^?9zE}@W^oc7Mo&z2wlO!f4| zVAA(tcwCMv=7))-4rlt-O*sLsdQqs`sGdtc)74#}Bg z-C_VG^u$yQN9`14iN&2o(fWMbUG|in!kjEqg7wqm`Zj(J(kJO*o-O4VRa&a>U21e2 zxbvOy>DNKDAt=pAhDLtL6_0z9a**hR{5HdL)73j7x#yrv4`t7p@~FsTu^_H7Ad{vZ zW|oM)Mva|_CWJphJ;FmIGG^;kDd+B6s{jIvpu>S}PvPOi*!YD4<$uP@+G69C)8pan z(N>wzwtI_X%tj%C7umK-ZLkCnquH=(3sp>03d9R_I_>z6)CO7 zSWNFy^L-<-IJTF*!=;WoN0ZPn{8hA6hK9w2)e9D(eA39;Wugu3 z+Q73%ZQdl!*%XqEbij$1lRUBAJN|(|onlDXvqN9^#9sO5P zc4?|i@Xf{L)sO-u+jn7$fVxLi57n6!(%W2+o%EkF5OqsS1BV-ul+{(VGq3~qi}?Xs z6A-wRS0Gmxc*=vBugmKnq5Lub^yh&+Otv^knzVJ(x<|vA_+oI8Jng)`Wku$w@`4*% z$?%a|Kj8u%Gg$rGM+{YiUSd|3-;}UFHY{i$VXGSb&Jt9rCgbVph#j$IC%3b1r=@)s zM(>hw8xVHM;!%N}&7-ER;8pPFMcOb?L%-14w-W1r_YBI;L+6h}M7aeJG)ZhtWsDSJ zM}`y`r*Ypxz`t4MQZ)ixa~)j031=|s+#{JSDTdRP9}krTGauX|%(6Epuae_mUNdVr zJu~4PabIZ*yQF-6rk^I6R!d8@D!-r8GFx7#ui?d;7bkkms;O#rRpWXvB}lKQyWNg$ zs59qV{!Ml6i}3F)VRg$I3b6$SUzLezUkaunDny{L{qxOU-SnJiEc;lIy)7D?TVkOC z#M=Uj3aW4G&beDLp()S9%;O{%JfOSE@MKL@?0y#z;yzKI9r9v|9^q7y(;!_Q@qlUF zzS;9c1-FeDai)p=m`m)p`3FX)h)+{AMSO~VOl^x6`F&yCI};a@gd@1MbkOGI@ zb><(jmunlcEZY#>Uo)6gMEOH=qZ>5KeHgS=2eBknJNSlA>0-&tE;yR+uGf~N5#NND zhcXADu8Q%^iB^!m1{Ypa&q1c9zjHXG6g_>+vLv4?O@SilS`!dDQ6vwzDv zDV2YUHXU{Xj|B75 zZ`p)G^h)H5be(PIj63X#02d^d$KN+&Vz+ybiL5W}XlKbiJ;%nup3u!A{6!xwvo&LK z8Lm(xqrrT|*!rZGGE>i4_@kLrHnWu1Pb@1QS!ME=!63_nI1!%o?15c8NV3U%9^_4k zdxuEP?GwLQm)8;0ir$0S3nE^H>Pn_p3JQ1q0tr40E?zixOAQ2O|r{=bW-=8X!IaUSz zO-}c&vkd8RR9sV@Oi+p;8&3M!-Fol+^4~O@g0-7z^$O1-Yuw+okej@6B|tvA&G5kc zJ(WnWYs-8r;aB{r`t4>ZXye7FZqE=BcuUBSiqN2n-Uu29nsP;D#tTLMHr#j5T>9T{ z2ba=Ng+o5*NO+ZfFB{QY;&f-L=J@_{Tru+IKAfGM-){G#yn@E*LBKkbp9!ZgH+fnx zhj2Od;|SoEeOr5FvYaAqLKvaY!dcd-A$FmJRl3B>BVNP-ujVJ#=cjTutD)%9dprL| z0sqhGMl0{0=lCQN(to7KIoR~_9QLrq=Mu?O&5K}C9I98Cw28sFxq1E-T(w>F*Kw^^XzLim9gEa-6X^}TL~D)U|+gW zyTn}`-i88>A@wYK!~MNPaZ2kD>LV?|4_sy)X0$4;v(-NMx_WxnEYyiF2pyueKZrqb zp?FZHK_50=rb(!1Pt~nS@fHXdmR(T)i?C*gl(*5MbfQ6&U#K`dxjl7;0>>U{e)e$31>$KfVkFPH6lCr7p8!7VfJ9_Qwmtz1(t&O5i&~0$DIZy+ zoTC2;S{A)J4zBr#6I#NGC@lC}iaBT}ZlmIzF0%XZGA_iD7{aJ`&%uv=wpNq4OL> zZ&6q6Rzl*s0tJsfUEa6n+HlJT7xS+=2b&F2X-3TLT@d=478*B;m{->43Cm5%7LpeHS2+xN}@|m5YSA z4miwaC|u_{N^I62U&2P6%Bg}k|9ndr-v;4wgOT3Q1dn(jHQ)kB92ufe&dO)9g-I9Q0HGQhhyHUykmH*#cq}g6Rt|utQwl;+NMg0n=e5iuTwPC>iy^uTQxFt)-e;=&9r*E z|CtwKLz0%P22UvtzK5;;Uom(~7p3c;HeAO%ABKOI0VCojp<)|*XW9VY@ZmLz5kYff zq&~rgRwRA^BvNXbaS1~@(PFfGjnUk1dvUSmvJot4$d%xmI;|||`z<{{2SOO$$W8|& z?rJL9&U*LBA}S~@AT+0gR@7}8Lt7@F(S0!($&_ZqF)LVz=Gq?othDooKn3MQQNar2 zg^aB#z17$J6Cbq$-;07RZtTAWnwjD(zm8jZf({wjbtLLTSfIcb2J#SAQ0=@EQRwhgwZ5Uf7fEgD7*)nq3-v8Ul zD*=^$dG(nC2N4m(GGE4$6a0hClnw%K`+m=<6J@^xegsGspI2@Q9Y# zw!_S?lFr2c71^dut5IF-G(U-Y&vzSB*V&+rY84U!Gkyik z>sYCU@*8u_(0d4&dh`a=f3*4b79C(L(n9Q9Xh#EtZthn*_W9fJ36h?q$~mXuu|1B| zN2U-YOD$6H>J|ly?fJ0QYcap_9}Uo~Qqo6jlq7hf3nJDP)HN$Lr%{|8c35!-pW7&F z6QcCqc_lG>m7a4-1K_p3gXBml*ABOSt;e|q_o}jY<>TBQiu^rM&_NF^oi^f+Z?nS0 zHI%;h;1z#c|LLCSK6$7nK+4m>uDVOST{`P>iD4N$=}t4t8gTzZYUQ?i=yv2f(wcP3 z=Q-(zLG^x^+~j<$DZMYeU~7bGkkDW5`N5^y7)c*c(Cmk#Ip0Rw9fA4e{C4RP4JviD z-oa8Xm`Jv_RZ0_`#2ET}XU<&IX=fKD^t#-C)QS}x|FbsF#&Ympx|e>hsy60bcyiSn zDBY$O_+11EAEyFbOV&{AuyCfzl7nRI6WtqtN*W1?32*>B;)Nu|>r*8bY1Pd5(?J)0L?Vtj?cYNe$CZtc&*5j^`(H19J#R@FumKA?Vdq?% zGYNzx^v$G$sc>hip4MEpF+P(sLTuYR@Xc(pn>0gz&P3|rB%2@M$X3Z#uq`f@HiZ@c zDTRk_0q#BtWf0`#Yu2K3zLyrigAG5MeCPc(9TeBj6YKhZ-(*g+`+u_lEDz}wt1RFO zq3_9JVhmt>1K^)USOE1V;4s^9AU*Os!WO1dZ<*6#SJ_CE+aMs{OPQuH!tOCt>>!MPMN~GmxO>Tl||H4N~=~ zUHXw+VE{t?Ef{GE9|Z%KOq(H4 z$+KFFx@8D@ZXZytfTk)OIfb3|3*q+BxIY@TeOb+-OtBM>3BiRpV0=N?NurU1s9RZeMJNYETpUzjF9{vng!}A{vg9MHT2h;zVCZ2kr;7(cp$> zzi-PZPQam4fR|lZC{KNxDzEeRe7kqBs}YbRYZvCHo2+k5Q|qjby=SSwM}7ul=IV;M9FWmSL(jGyC*^N zz8zAjj*)DQl`h)|X#RY*nfP!Ed5Lr;wmsO6BH9n3quM*lI{@XGB0?VW{)P@Nlk=vp zZeO)RVYfI5`{cI9pR{yG-RY3QBoCMOzJww}OM|&O{y~gjtvo_6R%6AzDi;*v^Bw9T zyBVtmLFQ&=3XB!kH=~Ituppdi=hBa}8{IqjFcrL*q zYC_!C_1{N%PZGYc&Zfj$mT!x`79ukegQ4^Km_fhH=PLg$Cz(3=1?;f1J1X5$^J#D~ zSB-6t-Hp?ym=Go;JNXxazYOI{0pU%+t6n3BL^hh^mQ}(5aQzl7^Y`gC2oqvB`8dX} zk+;8qyyDVpU~1O*7*J`1#$q3Ee8@bV%8(d#lP>p&)PlN^hD{CiArAr^8w=BKFnSJ} ziS~UdJLC|2d`_cX%iNE(1TS~$d^Kl)ewP?lyG?CXtY10vh%7_P+Mu?aoQVL6%|!pj zUkDv6Rkc1A`*x_u&{dnH0|h(=RF=8b*KC4J<5eTJG@^9aZvdK64k~NqL+kY;Lfon1e)TK zV0lA?IA!&4;Z=F51=MDKK}w1V*?0xMUI-}`Vpc=8?yJl}UWb{>urS{@Q$|)$uJ4ANzVetbtHpXeyNuB|-PB6z&D?VihW(k3RchJo zQv*UfmY>b$BHS3Jrr#$|%mTo`K++r7Yu#a{j!`fG^W6^$(|7LDnSZerE%vqi;XDQ3 z|6iBa0zeJ7VSFo@_|9(xqMd#D$OT)2^jJP@sM`y?ZXr^MRL{@|d zome>$?GWc&MDNS{PF)i}rXv~YTE-BZJgO75JjvO0k0<;6UlvQ#zt2$uMr|?tHjE44 zXH`cQMgaf~^mpfRBDo7_+OK?5DW@YvOkYP^?x}W8sm9f0wRCO`FQtMA7ttah#NdF)? z+`fw(8}D8NB3#3Ko^o0np&(Xhx?86q!jOzh6(w0-BslhdYkdcmZ{tA$j(J`}Cyr*< zwROK9F{@^#PlK`>uH|}TR9(h(@bM&(l#$^f$fA0@{G4K=491e z2#7*3^=`p1{l;^DX9zxRJ!>jTsbu7yGLYbY{&~_R`2BF+ z_UwLgxZm)3S!G}s$7}}Q>22j51wZzw=_(B%i@T9fe=-#$8b;@mFjW4LAsPhjj}l<4 zTw?-uzen$rxpk?4aw>e-Vqkr3_9b}2OcIKp!3ML^`xg0UgU1E~B(D<{cp$xJGXmr` zXO1$syw1p`@Q;Fw#L{0lv|XA=JEzD{7PAKb>p!1$3ulV<>WJ7SKk$oQv_i1hGk;5p zQo#c_C7Hf*auR|`yo@i3p9H3w!A9<*U5Qc7<{VLXm)13xsK|a`Ps}(MXd=gJCPH+r z2+e#h`y#Iw*mQvu#O*mimMM?+whD$!)Ka3vte^{3;cYL^bCH7%dSJHgHLwc`h3=|4 zVT^8W>-|Nv1_7V$&i2r{BYli7a!^Uau?H;a_O!5j!#;Xtq3;Md2`rHEAQyyp4cK6; zuf`T6VjObk8U`)gf(u6;ld$M!C%QpnyZE&JwqO9D1+^NwK|tOJHBE-E2M2)UKVZ~J zt?4MVSrU0_FCy_5F2tZ`p&=0~l0N(p|DF;=YIYh`_SJgKU|P+Z&E{=jQDhcr&rdX1 z=&FDlip-mL1HSi6sbuAQ8|7Z`B5(_9rJT?CiqUf~Mc2CY$;{KK^q28lj~Y@z3gA4t ze@27F_r#)t?DE*gyLZ08B&F`)t{Th9V~2m?2`%PrHjS1J)1|-G77Q1i4Hs8ePCu$8 zxq611H0pd1a+zbbjinD%?&k_bxDjD&7&;m+`my}h9Rvks-@Kcn;d<+qQ*GN>7%O7n zPtx_vgdpHt?5rv2>GN1_2odZdf91K9tvp~=;K#pL{x=DR+`t~vWfGL+o^PbP)<0)s z4_DVQ%gzTT+czTVpEe72+WS8Ce#bFqUXf4Q{J&Nj8jR?MsD~X7IbWtf8$&NNmFz9s z2X6Yo!JyYeK1vOIA)(3JA5+g&@bme{a$bvnQ6&w-Zq^kZiS{-i5`d2_CJpOJ+z{2C|mW=S;NWHdHaD+$FfG5?Bisf(4xfcoTkI}3$SiGRHo0{ z^`M5WfHQTjU+JeB6=k!9Gj3J-tb&Y%CV@38Uz*5I(@Xuj-c}+W8 zo1d@m4Xbf*$T;gP&EVml(Gs{)nt{AK-Nv4qledS6?8?pjz`sC<0f<6#ER zQI5tr*Q8QaXV?AjP)C&N$j!tT5mJrBp!cm??Xe-o7vOIWSpd(tgiFGna2CkJ-(`L}Ugp92?d))*HO(f_9XiU=-C$?DYf|fA947r; zRV%j&19oh1?ev(71GvkhtHwd7=_40{fiFygSF0dGLWm&X_7M4~$)A?<%D&BZYQ)>O z-Ql|@yUM$=)1_JssN2@S>EIHg!Mis!p1=_A}439y$;V z8Pf;)SG$PqV|5V*#eOip&Wn*4s0!Sl3$@~&v+4*1M?X(=Q)+R`(;|}7>P^7~FWm#Y zHnqzeQZDI_Ol%6LB<|OCEl-T&6Q5@frJYc^O|>6%G%`UXpRya>{c^=HOp$B!LYO_b zgO9rp^w62cfY83F5H&Kc_-N#CFHi*nqJuR)TE_vYMisG~7ycUVK;q#$0e z`0z)6_^#+A5JU>=vz6K3RO{OW6EM`M;Mk74CO5!xEQ0g&cqKi}P8+o)N$N5mNA2G( zl)^54!Y6n0=z&?d6Y&#OkZY!iOc=~O;}3R%hVyzvJr`p~$O}0sIGQ4}{oo_xzeW1O zAb(!1%NuO1@oRdtG1O%V7iOn85e^_Cgp#*=$;n~cVLO7tN9J9M9F{4SQ{(*$g4O~B zZ_+XBuB`QVjS?yz@zN5|eoJH1otjJ{a4iXzW!Eip zeD{oA=ziU`Rt$wFpSbj@{o1Yk%@t+6`dp_o2}dQ7IJ^Ig_glfAH9|RDTpYXMv0`5> zsD&I(bG=|#KJ5~hgT(72^slutj#{!hbS4c7hTbne#Vo3@x%6MMw0#w23xc>CVqHW3 z94}v^-&Q7jn44Vus1zTRIn>qhV@ytRwWHC}E!SB7FkPNX>N2NFiDnjM(Xx;7&TOM} zZu^M4R*-%tS9jUnKcRoY#Hw{}G@0*~Bn;^-2EyYK=UItpWVCYcQbJ zSWTOD%A0WG_HbhomEuJ%5-I0ih=`;Ic^)@ZS4^sdG-CbQ6e}I_MI@&l`oR^kNk^0aLIE$EPp;J7>`e?$9KT~D}02a8}zHV<{11W7D(y0 zl|U!)vrS)=txn7Q>H-PV)mCal;ltZep8Jmw)me<2?wV%UU)))W>|*>qB=c^A!Z25+h1f^%M%<40u}i z9552Vie1GzAa~>av5>8Q9?lS(%2%;z!+nFG;I;6xsQzW7whF{LAxKIq$s=x5L&YSE?5~+-x+&7v^ta-6Cr zn5MIklu-@s;_uT2E``!g$J$2`OZ=A(`8o#OO z*ZxqFWp6CH74T9Zu7jEmRa$|3^liWFjgi_Ta$Jtt&ac$~t`f3^XRVGqC?JrP-QzH| zBP%d?wE;#Zmo=Ki(uvYutU)gt#u58HNdI6TWcU5#%~~Fp@a96-S1!=Mb5fJ9C}?2K z5S|zXqAEtXX8)5)v0{o*#*D?p+O2JMT^O3WXq4>*7-u*IXDwg+^%eRwS0hzpoUqd$>wz$)ZH8O6@b| zdL27&YtGQV=`!)&BXu48Dr@fdR~Go?C)Tb6H0~Ch99{6ipIeU)uD}jJ6wf;*i(+Dk zoz8CL6bMvy4qGH9Z#Ey#d9eKW&~vfjV~^qV>(%W#0)u8N*YA;1$at)*`+jAW6eXoV zy{h8^Wt+V!gbpZudR$JTh%#*TE5FrxuApDV0NOC#2M?!g!JVdYS#pGI(3W)0D?_uU z9EOpzB=kntgF++k$x*Hz*bME`<(vDPB(-*S1j2I{kSlNxfaE`&Fr~Gv8=g7Xjwx3& zWtW|Jsmtr4LIp3qh4|oJYnLOGuNAFJuL|QNj(fCxS~op9b3E*3iXv=hEgg5MAtudc z#htpssifVi3Y1j+oKgoPyqi=5A6AoP~PVOce9mXT~P;(MhCu&=> zYW(x4_#FLmxdk1)MlH{Ct){c3>Gp63hs%Z4Z1y=#uGbFMVwb+A1XtzB&9KiMOynBh z>c^SAonL>`#w=LU*7etcV8IwZfX4cS0Njb9C3fU;-MGoqEiW&;Ux@n867IgZ(qT5l z>v{M)to)*QkeF#gYa5j^|J@h$xvs5oB%aZ_As@zVyGq(YusyDtqdOH-%8-f7Z=RX) zsmie%U()mLDiXVD)!);yp7B-Z`g5p~V=V#}pVUx)IcE*<_xQp+V@zirA#ke^=dpf- zPRPViXu~b*#CNdoXXwS@6Z$sV=Ao*|l58!1A!p*(bZMbA!)aM8B}P6Kud^upAY1H_ zOK#KF^Jwe^!I-~PB$&5A8OTM~lfMN0S5I_wbc9y;$lB~!Vf6`#niC6*p<7T8Y1Oxa z3xPv%6uAF+VaPb_UWoLnz6s&C&X-9IM1CSpS!}?@hTSvAxYh5*iamW2xI>=iU<8oK z&V*)VyH0h!!N$gXR~BBEPCP7j5go8Sfy}XzQ{(w2xl73%q<=9Xk~9$!otTiFfQt>g z(3FtaJZH_v#E zee8B@sE@n=y}qPBduVB`y|fCC9%&&&+EYkolicEO)Mo*v9x(=>^Ty$&(9+$!D-c=W zVD{FzCk&By++~-BE2;g(In9d2_*0tVlm7Aj=yL%U`zxnGZ-0hs4+HCx1Mi* zcc>OW$KEND4n}sbwc=7suAR;IblbAZ_SX`{DyOo!TXcWi)hVyVJNp#i{pW3%jc%34 zAht*Q)n3t>a3B1lIL-dv<0B@&l|t{lNzks;uUDDzjTMqxENU4XrvHfUlhL%7mzSBt zrM99Vd+b;-cz_*DA)f^f;+a;v<^E*8*-~X!l}vt~Av9@ek3zK+2oegQJFPBZ3(f7`@8ts#< z(Lrp^8prJsDV881zfMvwFj4XkYA+I7R3lGs5mZOHqJluN^x7TPJK8gexbOXTu*76c zBZJQ{O#tzb%3?G!%nul^DGpJaAv#@1$d@(%`P{ZpduCzuxrch0}QLHMzT0f|@UH)UUjfkTW07FKYIGbcLu0RDMa zZBJt27Kg5ZG=hLYPRM$?H->`82s_83#;jP;a+&ywyj}(w;qQQ&xOw`(gYji(2aT&bu$@JP{A%ZU2j=vkYtcefzMKlynIQ1CeeeB=?gpB~&`3 zq@)ETCrAm>-QC@c0VAZ5l#bEe;l>!-v;T|d{f-?wc6@j5y07c{oad$WAUMq;Xz5Q{ zBfbTOJNsfG9fOVI>~%k2v!W!UYC6*Er+=F(4w{C^9)co6d(BKEV(&H7>-$09+5TWX4)wd6C{-B~zJF)*V&mO&tdY16>6T zwbkxujniiwoW1evFz{#!U)7(a@emR>+tJi{)a#xu_&-Pkb3HbV+<}Al7;IJ&V3@yl z)_%pyTb2Hk+EKz8%l#&y5Pu^i`~7)kt#{B<<iubyi}*RKWEz2 zTfe!0;A2rA2a9CXiThCJ3+KG=t5XwnQl%P3Cw}u3nqBL5B!>&D8+jeDM`Tu2rImE9 zYDZ z?zlZ-o_qz@Q6dM3_RXu)>M&@qXQZU;(nka455{xp1x*}Hy%TSj7DUm_5mSEH4jtDM z1$UDweX{pa`NN9AQ8Qo@`0N)`;N68a&F4>>IFmsrcaV?GLalLC#M0uQbn#i_U7@Lb z$^uiw%T%dQT+4*y1<`b{|KTcNKs6HRd3oyIMGWdVRrh=E)lf!g45 zZ${?RbKQ}QwQjOVpkicJw8tzUcmR%?UcGxF%_=3*bG|423(&+hDUXPM;9 zzs@(QUuaffo~S1jD!)k1+e~$cITzSP#Yp~>AP}xyVzU25nJGy1=NsB;cV1&MNj~mo zI})_Z9j&sL}y5xDl-oo2koM7O34;;_znxzDKc zP7HWioTOaU;__=1eAlvLEvdpxa0X5jKau5~862x~@;+wJP)7k+ z{;mHvxgdO7ZCG{d-T*Fp1Dz~y^&J*=&J9+gVJ7zmi&CO#g+oJu8fCX&$`vgpw zqtkQJ;Y4bnHJTd%@LlXdp7|o?iOpX6$V}X8%hu)|g=-|DaoV(`;tu8qfjyHWBMDb% zzohi>|p6RAi&R>z0r#(p*z(WZxzogrDc7LAM^PlnHF-44)O5xYYaBn=i z7}PXp@}^{RlfBaQDw*9G2|#|^H5l+9iY^||SLm2sI@ISI0quE230DwxmD$quAvF*F zIsj9IkGPJP;5YBrpEnh)uYE)RKDP+zzvp)OcDL<2Ec0Zut*0&OGK{`D!cpKT-}nzH zu&0swh(ZmmkFuVkA0Y#j5Hf#B6)-wV6*RAGL6~Xlq?GIoZWv52gO)4G2FtUn;d1pd zn7(U=p1uTC)kY$INraG4f7JT(kU?!@MKPTvip^} zg+)8IU5Pf>L=VCHAXl|xo7D3{bm5!+n26~p==t05XTAr;D__ydJ0o57wr4Z9+-4w< z#U#{_k$OJtNmIn_M`7~T524hSR0S`Djqk#gmUvduuD#S6hk{I{q9P$%FZdNl{zO1k zSLD8wVNTm%mI^0Z%U$EGeTCEkY#S`Vpb+9W#+p-u^EJ5y+&)#ma-QC)>rE+d{1iQp4 ze^+Ht#r{S|4(xkquV8Zgl;N|zn3WD#F@pRIusq`5j$&Y#jkUkx(d#xfT6m{W9?K## zO0uc?KIGG!8{EtX+_l`&7AI9LNBLbebGGzt1IyfLv5s;5hMiI2-KMyN#g2ho#0#GHy;1N2TUQ7MO8^K7PeA#! zxTIPeLna=?$7O!PJu%F7kX8@ZSuG)p`llC9^#0D(m>JX@m=gAGv0WTBet&(|zFqhO zbUR%sipBLnM;`tdoUt1jeqF9h8nv_kp;^@1_}?2!Lv;Xy_#8Y3n`3tOZZ&ycz?3SI zy{2a()7x;iiFZbM$5C3dHK5tSkn7(&lPL*P#yw}MRo~`Xrd#`s;86sL-1d5UQ^A`y ztSWvBp?;L?Fv$nBBhCAhmHkh<(p0rSI!4-QI1PD9#n33+&DZ$&oRBLzDDYH*wqN8| zvs=ofYEl4i>Ww&hoLFq+rnDf{?ym-8U1{Gq4a^37f0r#bolNqqdt4}j@~Jhgnjy<+ zf!S(1qZ3w7cD^p@h?Ir^_j-}k0LwPfSzQ52ov*}H_2-X!YQLi|nCZsSqZm7H%IEAxCcC*j@Vip{f+DqV4bntcP*iMMscKpCzg+{2C- zkqb$UHy?=g>#e@K68)k<*xA~uAEsOPeWk56Tqt;=qCRdUYTNQz((&&U3uo^G5wDEC zkcETS=;%!7C87Kzv4E;u^bqp#lcx_H$E+3q zCx6y%&EhO%(XHi+GzbNLhSK(NMnO`yyl6uVDVCP=17_TB2_$$jG*g8f==rhOysPcH z>{31Nmf5DW8BJmFAgePfTr`ilioG+RV3^IvnwoZ#RU?fNjpc+Wr?v{c7b(+B;k($V zbZmWhqq#gcX=ct}iB#^TrI@!%25+0~mkLYdxb+V@nJ8419zT7fCuvB|ITKgt#mL*@ zD^zf^>HCJ&iA!JgLgC|OZ)3@r?UJaXCq|@6;p?gTwDFT)w@#(L@1mII1n@ zpJVK`5~^GW);1}3{HTax6~>HuppYFEMEEuw3xx@Tz!!vOVZi4Y#&s&ur=UR=4&tWT zsXAVwufdq%q;+Ehh>lu8WVMbz^QCc1A77}n)7slE*DH6k25<3qe*@q@gE6vjzjpcP zc}Y5XuLC{V5JwS5BdI95ZRT+=K$%=i}B!g zoMoCkRQD^j4PTQc{vl#DZ=u(R*e}5p6DOVAjloC55w4oEfRgKT#%{8(n8%`@e_>kk zUXQBv`rZ*(aAa(c-ANa(L~`>*F#RkpKcx-Q0KMjNQ;YO7*SI1r2~bOp(F>5Wh$v3R zC>7yf;to)&AEI8Zt1kd1nS<Xrqzf=GJV7w>^G9IYeHE);Rs?|tLI$s8Dc;iw_2LmgZsCjqhf=6MOj1Q~0AvALR3&%q#+8`o}q*(!~Ev-{V?}hcNRT5Cjpm zX=yD!`WtQe$Cb%tWi;K0wH*2t6+W)^crE$>Jd=Llx>uXh zZB87#&Y2F`!mFYSedyZtJ7NS6^*%6o8APQmQH^A9_3X>y3COvD!tR9x$o9rFggeID zpbOFb5>Ch9oJR;z?o+rtZGS3&> z{091Co$S>PFkv5HntyL7N;|!LeOQ*~gxZyN#5~KA9M6#|?8W`#J0WI?j$sDQuL_Eu zy93~1)vsQUg#=XN2F+!6prZKKplGicBTVy}Cxy=sTA%j8{zS*YX2x*sgB3^Ob|VOO zWsdg^`m3aRAI;v=>)b~elPQtHry1Bb^t~QEVGOfmg0iIyZwb`PM_m_%^}Rs?TbkgU zM!Tz*=Q4jD<_aQM3tt7S`}Esv28%O|R{9o-UXvonU!i2r#hhBRjQ(5m7!&+laPebL zbP}+PP$%i^;Z{@PdY5fauf;>ITUx&dVuRa{;|&6;I4?duDBujBtOL`Fq-5>ppJlB; zP8y?E8|606Qd(G@db5|VzpSJ9{IdPeq)G)`3jg7u`HTNrqA)l6+s+S)V9~(k57Q38 zF^q_6q;1K>mr6mW47d%!RZF!WIn4y!D!Q>R2Nj{dC{JfD~D^qORG}?0D@`vg1W_Q$WN;j6UI$9m*#w(#85VlJ*d5K+I zU3)WDGM3st@4+9@NY3f@(dh29$al8NK-frg?U>7cwpS~+*BIpC@vlvs3GU4Eb=k4$ zx78o7T;=vw#Z3wfP}s-QP{eN;pQPfIZu12URE&t3$7tDH!$8$F zvMS$^#|6McTe4d&SkJwNL2~(gMKC;EJN{{@(ECp^@#C<#!YGdeG`w>lAJ=l_7ZecY z{K$fh?AjjH3#(e$RmCZBPoJ*?gZOpJgWiOUx`Zdmtt{2L+MoWSkOr-$FhiC|D%$9U=Vw=C<~4 zzEs~3x99e89{vjf>MLOH`zu}Snidnf_a{xl&yiJ4S^t3qntk^g3WEL6fNkP=Q(26A z8SV-hdOnK<+Z2CUR5bl6>YBsYSpREG;z`6^9`P?VB8|ya9$m@lhuAArj7;`cUM#f1 zD`cU_eiTETX&+)&8X%3K`K`79j7+G-v_^fIoP=6Y)890jFWWm%50RREBm0yX!GLS= z()83!$eiM(?Nc9&2f$e0yjdO*r-kV|uR^X>wL5N}JT^V1tR_1i-|H8yWZ8S)B4fv3 zWOD*JUyo&}GW7gZ&}%s?B_*|YMd+gl%tg8W@kYb?Rv+OdYadG6EOHbv5N+;Q`!*BE96^-5$-gT?5=U!$dCXs6{w*V4jn5ZzEfj zPHR(W|JChGH19!i#`@novUHZIVwki0JXb|*r$n?Tjn{te%PNqw1-1sY+c<7h! z9vc}M?adIeOSz;p-dKJoz&4|KSpQ8PD$H_~k4;QSXlC6}uWvdrq{-ueLMy}_=e=t< zU#q0Yc&U*RpDuAe`GeRvc(J*4WU-(|)cu#GGH5~1*M7a@t<&k^`XAa?{zO7&<2q=O zpqAz=NUryo0X>U6@<4HOc5&IHCNB6+@7qRrx%#$1Qzj#Su(RT{Bjy1v*}a8&GU}O1 zj03mzI{w>~v-^wN4pcA0%wTm|^1Vpd*|s-WeD)=8*lT!c^59drZadEpkXwt+M_DsX z*_P;lbGzv76#kSpmV0coSQJ6_Tk)^uAbUGL1u~ersBm?-AsYGvb(tnOX2ja$CEtf5 zIyt`Z)LR#$OVlTmU5B&gKCR4rvH>$w7}xVS4+Wyh@6Y(1;+{XIP$oybf81sqFefKy zF|)K}bk(#r>b_2J{Rqc5b-7QRI#sRR6<0W!>b(25)#%!d701&aVDy~rMJ->Nk6Fq0 z4FRUmCjaKO*$J@aKD^p!*0!Z35C2WfA7`2e8wTkp3R83T@0_1XZn4Uh3cx*tZuxQ| z+fEhOK9^FTBL4`J|AV4cigY_Ac_Y=Y9nt&1jEiwB1aa30pVUWSG*SgV+ty3qfb1#I zZ~w;Ok-wf2u%P?^c)t0XM02p@=eM+mPl8fg!yHZ+u(f2G%@SZVvZ*eHuZvCgnPg$n=$2jcY>A%J%W%2}Z zvU$pC8lxqn&n<(pmJd~e>sIKdZh_?IlmYcN8v;LnkV_AtF$H zW8OA)lv`0%!1bEjV|h@Hj8RrRUG_SWjghE<>!2N(g^_=?G}JfJk{fXm&VL`zn;Ld* zG8L7(y-|a`8zr^3`xYhfS^vdA~@<8n+FMx+fJidaS2 zvB>0~0lnXhAV+d*qdwz+lZK&XDFL45*Cr|v!3VxNuvNg-uO8zFaLCrv9o?62ERuO{ z_7^M+9M&rBj>S?9y;{&>bJy_=IU?TKEGnR#1T$w;lUQ1ea0}O^Y(ZQ#X|?~{UmpP* z$NBQoCXcA$5RXXOje907`v_8HT8t{dvza$IAFvNf<{R*=igc^{Jh@S!H{QHfR73mJ zd5vn>Va&KYTHy8OpF}2#+`J89a-OYJC+F%S^m&=oglUDYqDEq%OnD7o>7(9EO4goT zWAE17TVsokuZLF0!X(ENdQqOSsf$)QI)>-KFZMkxFQWD~Sfje>Ma#1=XzO!Vg`VYK z(x)qgCEtybsn=X6E_{80Rbl^g-bO%E{Ck9$V8&D*|6Z4mZ9e>+CCyIO`wa_@9Rf&K zzT4SX^GDXOXb_P6M~c7PO@6AQa{HIQ!u##cdE`aid0-V}x7D8vF!5xRIwo)tJqUXr zaC`}5&O^yopG?SD*jYQEHgiSLD%|O8B=w1+VNgQ{0X?ek5k3A!@4FUYUi(nzCepc; zLq@laH-xG!3P045nwHNQz7zGhD;w$_yWsPn3m#A#O^iqK;Lbpm`8u9hhKiUdW0NX- zR-6L+ag6f#)j6KCtQ8(1eeZm`*HF9sAcnU^{L7-7|32|#*^2<=-J*3IcxYFH0ba_} z=l9cEoZ96-qyU{820u$%-%;MO!JC_+ZlMSFf)UH(PSBcrmB3&j8msgA)4GXA zAfX>CVM|s-);~!pCi$>H24puZ>vhHYZ|EewrY=- zW`ind=Ate~NI~Oq*2>$er-9=WOu8|sL7)q72vij5>VWd0MYwifhF>B?mp_C(>+{4@ zy5XQopt%}kug1vt*6IIF+CD!|DSI5y;ksloLMEunJ6!|a<_ZJkp6l>4>LH_ceik_n%xOait48SH7oeN25lwP3;Xy~mpX67y}!^BOC=n%=F5qfwnV+d+?h-AwS z{ZQ+WXC#SU@WzY5Ix$cDFn(0sHxQ0vRKve&2 z);(l8)h2jY?<-h<79`O;8Lu&TX;IjTwEh>9YamI%b{ueh&2L4}<37bYaO+>V|7l z56jm3=m&9NTQ&SI$#InFYk-@doLKvd!zfwDF7l?ey->!r9?RcD6gWfT?a~;C^r4s{fWOV~k_#ob7Mg@M63Ki~B;Q;6Q?{slt;)Ri>k*$Np)tn&{cwgwdjpe{ z!v3r>&zqD&^I}&66HtQEkGjqBJRL(Xm&AdSEZ1%q5@fD{xozg7&>EskUCLFyg4ca0 z%0h*pN$^_9vx#Cqe)A!XtAfL!o_^8)MUCMo0e!Rj-MDeUhdcZPbBj)@}nGU>?Zj+FQC#h zq5lm;+GK=$R#&|L(&M+k*0vC=4NAHJm9udQX(JMRD zqCcRH{h-pHBV>6E?T4jVUu5fpeiuprQP+Kw(NIezcHno`ufHlH;`(yQ7;r$zn9!2t zPL*L-)Pl+l#RZT>>_283a!>BDtKH2A)_<9lFXx=lu0%mWi1)8TLwbaXFn2*yjy=iU zyV{sXS+G+a`PAtzkq1@RuiYaudpVV2hUJ%3IcR%yr%ks3wh0Y*HJ}v@bW;rsqlX9% zG~&pBkPx%Nr+4+*h>o*zHhj!7Iwpb`G>e zx5|>Yt9uvON`08J2yH#=H;1@65lvOGG7Y?^BzJkW^x{zA8|Bl|49)-IXro<;Q8lM8 zieZug)uyj)m^4!bdqK_qmq6j^;Hj2o%}%ZG^y6))LNRck;2F=@<86B88GNWK0pe0? zIr+{Y5`m#doJfNXdLza)*j^dRr)JYxqTN~hb3ERM9#`SxRrHnC0SBTP=5V&1PqngN zoD9CrOcysWM<_>2HK|kzMAtu+pet%X3nK21)YUR_U%VI0tp$c*Pz{@!dVVX@(DI3G za3o%_Ziv0R*1yX7L4YPTIeeg!Cv-Jl!SVO}^yyxoaOudbxe#OYNEBbFU<29&NZO8e z7n=P+As_E_y4jgp=n9Ck(5n}Zv`NHV1aEDzKm8vRMY7Ri zaRuuaXKWV{M<*$)c4nUrZW5iU^NH9kO(vnIWZlGe$_zlX=_Y}~Zz@@kIJ%VL=OC=n ztKcN16u+Ii^_9dohWSF^ITU;$RF3E?`qWiZYp=;-9Y} zJy?(ojc2^z2S0NriJ2}^b$TCPEnWyGB!8vK`@?BAn7Y*`RUstE7bS-w4@5V{RZSJr9lAxTPo4jg@p4ADYxwy*6ZfN4jzMd=H}M9?et**xPl&cX0^n42 z{x9dzRdqd-cr!VUc?-aFUNpEg>|kK3c@=iR!)DJi!_8f_J^Pm&-91v1;?GB4(8Qiapnxv*D7k&E3b<$p%H3RIWYUbf#n9hHHW*> zC(8RZ_XWT@W6`6}Khwn6_uea5^(;pDqSPA|GCD35lthbCmA6_^{y}H03c6QT=_Bo& zd$ra+rX<8c-?RJd#AsVZ9^BuCFelY(VuyG@h`audW281IV15V`Mx1xei_$dy-r==5 z^vafI&%$yGW3sW0OGIEq2G(q>r$Sfp`@HraUH}bT8KS;Pdp^`0$Hf?}CO)SMFcrMp zv#u6vR^D;Z0<;q;Jh|EH*WO|~5Dkbnxn@M@wx+0WD#pMV9;6hDDpoH)h-7Ji?x&x5xWC; zWw8fZ8MnZkQ=*vmU=4=}J$n8b26X|WMDa^-H=8ze4+7s#bhZ*GOsG@i+#;A5_mKUmn9!H?Yq)X%;#h z^%wzK>4{72u>XdtEM<;1v>R!l^pYFB<^-Qj9p~M`wuhS4F5~PIVxdF{SD@cy|B@y! zk(^Gf%}9q6T`nWoX$e`u%eeF$3B6!VsG%V-Jskf_NjTuvX58fT(cn5-O#Q|SIK z*CSULW%nN{JtlL1h$}%}U>n)oLbu6m<7HEWTG>`i)6qd_rV7ll??R1X+i2J4Xq+HZ z5Y0x`r~fX>nHOLh7Y4@ylmU zR6oyPigN8k2qS$POp2Q5w~+Jp0cpZ0A~zv~=2v~(KDyN}HS=I$`hbN1(^IDIt)zIn zu2>|iA}Z!$8|fJE{(|u5023i%p-GZ_<9 zS*wK#n3p%H7H-H<$v(gt%8;nk3rGwH42KoR~ z5qw;_$7oG)q3pK`iyu&u_@J73!mQ)LJTMVp5}LxV-$6G>`nX$X({e?0diKZy4O z+Byg;;kH$ZthYLnps=9(hp3xh^fZ6ME``ok#Ix8)486^Hx!Q#GoTr=wvgju{q5IRa zu?y`A*3##8N@%R+YR5C2jJ;<+yLxU=QsYC$$Eu)dqj%M0F2d>m&jMg1na>x}HG_4} zkcC3=j^Pa1uM5qC6;vTX9(dNT4;>?i zfJd(o@vLE|4yb}kg6J214%1q&?QGdylHZ_K6=j=XdM40)nTnt$-dR~_gq=Gc&u!SZ zCoP!~=lmjQ{OzRM3jm4Ca%(!p%M5E{}0cc15OMb92WVP%K{! ztS=qszF2SRf;H?M=YZZ=W&qG8iD;~N(!iFCX+%z66Rg@PXn}NPVsyzA*bw_OyU=l#E&sA zAD+ZSRl4L#2RLvUH)kFv7Z`a|yA*!SZIvLgKzLrjPjVYD`_nRExHkX*4R^VA=pTw=z<&WhkMBTcXCUAawNH5IMa61 zXlgw9jc{vViP^t3C^rK%1B$y^6{ zH7VVXKcRmwd4Va71s$9sFEHv6KjqPgZG8bLG?sJTA_HIbT${1g+%dTqgcp!s3JfDW z+dcE-DZtiV2cm>I=I(cNXDJ5e?r}#K6Qn|&|H-#uOCDgk01h*^NvPE=dp zP_cOz_)}5e^$guEVxtSZ#a7G~7#{i!RnbH%O7tKY(|u#%(|gCDGTN|u!JvtFh$aVt zj!)aLm$E?GN?IcM?RKKgE^ezr7cPzJe8hg`$Nx}y^EgegK#ZWy%Q+5O#lF&>{S1$P zOLm%+;!;b6u6)x%5!+k1yLQQhVc&sKF)}Vt#p9J6Dghvx$29i@UL0eF-Gu2NFurH? zJ7AYoA{6Q8M>qgvTz>iArKFvF_o;ezoAgS3@r^8UXnLA0p4oP96S^zSzZ*RXr94NE z;oD>79X>xEQ+O|!fpYFoW+$J}d!oV^EW{Y1t%rWK_;!k-{WZFKv=gV!XM6K}JLmC$ zFI2;@2SM`M(*SiF|s3 zhuYPXf~a@%SqzG!gJhMYl?8{&dYuZPx8l#mLWaM$1|oI}pr`hPBk>cik=%tV`~jg; zJTY7&nV;eF5C}Kxd6hZM^A}g+(D;)v9f2wScR#DR^km=GsB4>zX-@sTe`|!UjUMLo7>%jvQbON8w6QheT7kT?iE3Rr(MBDKH?ykV`Y)1aqp^zp&Q&Og+z$8}Le^Nd=?;m4$ceC69ki8_^M2kMbwn%y`ly(vZ^Ofh$N=0>dr&)jd{^l|9(gM z&FnB5P|sd6Q2i1jcHAv^KZ+62VAHr6lX;^hJ)qcHl9+I15mF@k@oGk`LVs|ErsO(V z^NvD+pil|Qi(%CEohLR)Q@e7NO3etsMT8tr%-Zz+XQjcpIL+|0R-cRZ1N;4VtqF9z z>YWltYW2O1cd4a?NK%{BMRdAmiJ8yJ+j+7eJ3cpw#{dKEv6feh=R`1Bgs3^eZ>UWL zwe;~P?FhHIh#FGNkSOqIo8;A9ZtaH(z2(66VY7m72%@M<-QNdKA8Q~2Yw)W0^)Ei7 zymqeaJrQ-9?_yb05XZt$E+cN@AQpu0UEQ~z)ZUR>B*eB-@mWnOA^D*yiVKJPK#XsK zyh>&HL*FJ4KT6wS!TcG?Yiyk(nj`3T@rDpgaLv$RVZ)MSF|3K{M7WYaST9Fv{H#6R zm<+E~jbok9d4VdGJ}>%^q4)2>x7d{LS#dW5e$b`W!0tq8O_q#Z?EO6qb0*ZiW5wcU z4r@vAuL;H@T%NGp@D$li(w5^aIK^$lE>3|n#T`2;J;G-5E%ELq?t0d!Jmig;QTN;2x~B)tTIl@%f~q;?6-m!39QDF3zrx z(EAnvXOErisws{mEObycEWMBME*LR=Vh|vT;CM~D1F2e>z_LsVS zYeIx9r(D{Z)YBTf;rGZ4>7>xAk+mz^e)PlZN<*B_pTxk|) z+Me!0Pk;W94m%|Xf^p})S{C>#!-#@mlC({2b80@lLF`Im0!nK)00bOA&S-XsjF7hk zuYB(&fqaG;I&L!Ifa)MwkP-*KobtCszhB^xDuYk>(Ue~lzu7(c8}qYNY+L4H0bolx z@m_vv%G70NX}N4AY7VuY&`(p4Ry z@?qLaaXXIz`|hj|ExW9Gs?9Z@|JkeG`s4*kEP5TSs#|a0HH+K{3!}&`Vduo%Ae}{` zFv}}cmb%98--{&mEfg|nq}{h~%>s_?q9BeZo|s~UV%!twnM>{ZcS8E2OauxQI!H85 z3GSOnt}Wu{#+7F@VB!vBNJeKehPqgDinq{$Et#YQMTmWc9@-jBH#s0G^7DXZQZwms zQhe2iy+IK1au;_EM(r0VzTb*4L%K+3``u6GOd$mmJ}YHKEB0Q{T}j$nf833`*}wCx z?t7O=rv-2AF14T|^f>?|xu9|KJWs-_19H(vYDR_C_T-FX9_pASKmU@sDlb~w=>4%8 zLO}ZBgm+h{1MxhtO{p%N3f(^9XN;Sfz!46CoOi4QKCHF5gq!O}P%yj5%mdmnIK30@ zs?U5wu?a$E(>NEo)d)b~Cqjx=$LT&6pVdp1tbNWGfqD?bl{TmSK_YKLVoDh1f`ev! zB3o+KGIN4_WlR{d`ZimnHxTncDe2t7P&ttyvtY``MpSE7)+%Gt$EU5u(b4tVGkpu2 zkl(L;SZF!BdF9vMussA;_VDk+iApmK_!;;_Ftw-Abg8k`6B-)YKlp@+!n+Zkluv^E z^;BK~(m5%)dF=xW+8>Nz*&nO|!h+y2$k$7OTz;ujIAE-n#AE20|E~ag3qcJ0?Tdp| z7z+rMw{~LGw2Fn(e8Qdj@flEThRTD?QE!1BZWqp6_p?vW(p;_kYX4+{$84y{F@bEy z!Rkp|;E`pJ=k-wEJa0IlRe+_xDrrna$rc-0!B+Erd&^aV6i~uu6F}%Nz|Ct%m;G zUc*9iH!IcE;$9v^OanXKvB-c^A#J1Ln6s&Of#lyY|5jYd&cv~EmbprO&ug)D)t{1C z=gX^I$kD3WU*tF{Q>-0PPEWl(>rBB4Umpg|f(C3q>sgrFeF?aplRV4Y+BVtY%U1w< z?0J-G>{faQjYaBm4D|b`Z(eIJyyhQ^S$NuN-;gC=2x>L6=V8Zs{|NfM=Bt-#J>S{+ zK)jPA6Y86ryq~_yeyuAr2THZ7XsN0V>PB5@HtHbSAa|)*Uj+UN`s^Lr9pKfLn{-%Z zlyEFo3IHbBA2Z$d+edg1_989{6xVim->+fgQ1<&j7j^8vqzamR?#mJGM(B;U@$W=g zwjcZb)jdNKo*679t?C5kOD3BTe#ijvZw|GG@zGA_9hD z?+30+LPrY?P57dJFzFTFyzYt^ERM|Kt=EnSY}VJJnT+$wk=bw07LxkRGGN60z7)N~ zrDo`#2^@5wZEX~-SIU8BfF#rgac?8T0tOyqL zjz`$3!}IfHv3f!l7nn}B07H!~`CP5pLKg~}#a57X{77#9_BPvOOAOFo?@-Tfy5O_d_s@@ zbv+>?RAs67Gb^ecdkGt$<>J&xcFVK3Lo^IMJB~eCp1U}|)&WFd&Ob{~m^Zj@iF_0= z#V(_L@`Miog1(-2VQK}2_DIkH@U~Zzi37cCMdGy=nQA)d1CF?gywg?Mm}#2E&qx>u zvrj*swckCD^j>x-V+!!Lw2~WU-{8ULdYdWW)DBs-%{&IlZ%$ITOM2Hq(cxLId}1## zKS6g@yv6dU4aHL%%(}hLx-Ub!D-=B+h#FC?O`UfL7%4FPjy42;{adi7YA^T0&aPb8 zQ2i0ehu4JFcZX{Gqj%)HdE>>FW8Kdw@Wfo#hujmp{YADLX{-tf^=fop&f1^FZDkfK zz-W)hKYK08!XQcO6 z&>=V=uAKWr{?xy4XiRX2Atlo85d7DIBq2Ln=-<03W%}oFK<~V7>6gxttx4e0!TU_7h-b5k(0WhvltCAGd!& z?)kHMmj1Hr-mi`5sr#5MZ~gG+>IhkwuSKBZWUYy-iTz~4o8?bE5LqYLRL;9z!czD5 z;g+VNe)P8uA4xGnV@+$3p8^Y19?LsZZk+?|spnvOSVZHtD_%jv`T>G_!eDbD(e*fO{^VrJX*2iEl1GjF{HcLvObM z`tb5dZLw}LQZt@jCd$U`$0_lB+J|{p7U{rD%0raqLf+Sh7yuRcc^C?HsTY{55&GM6 zGVx_q;^aQlD7E&J#5gYM#Jkk-Fx5|`<_O9SVtYX8XmbD^uNFeQWeln}pmEmeDePuD zmh8^*@AOk&GvGzKN`0f#HN8tlXxRJGccKWQT;Dqs#9LiE;|fbf3gIWu&Oyw+;H7w} z&r#5nViOYUCM3gef_r{REt2K~aVR?xBQ~Zo25<$-^_5}`O5g{`*9877inTZ(YD2?; z`|e;Gxai9qPf(z(O#B9W%vu>_26>+xkjo?3u&^egAF#rYq<0CJfn)}sVLAGc#iBU* zvf1r9hL&|r?}o8YT(~i7uxbCOWfx~ zXfTlC;St7>y8bG}7dy@Kj(8T3?nPWh>aEx2+!a9_2t;03`xwJ#O~>>F*Y@)Q_cYRk zPRXSapLppe!Uag2b|naUPo72yZ*_PbXU;q-xP^PnDLneW)XP!nwBPP zn2}nIA!ffS>A7Pee$RPUK9ZLBeVL}an0ueiLXp4EE##N~{%M=pDO1afY}c|N3!Z^R z!-XzYpPaz@nNT#RSWwPD6(P&le3J&pFbMJptCi#}3|GT}H`@QY3d=#;{w|~AOKww? zhGNcS+OWqx1OvIitO_A(44~b#AC}jI2vmL`$xLf9y;cn@f-IvRT?lZP>PQF}b$+Oq zW(EY7fwAt*GB5+DVIC)GJT*3);!ZQD{NYkDI0SE(VH5zE@9_3Nb8x}SP0;f=iL=%c zrMa&AewR`fc}iJV*uU)l=Vl_y6@o3$=MDeD{}@VvCPtbj<@VpDupMqiV-719s)c$@ zVj-qcX>eE|jFtu@cc}#bC*z9EhNd~#<+_JgnN_tqVU=N*>|cAu2Ldkpu?z}F$hHw^ z!(TvJ=b2!3TTkHiAIT$se*7Rs%*9bl)4dpEQmM8D}-_ zOxS{k+N+ai#QFKh=KeBy9`2?^pAUh3N%t~jnz8woo_Q#;GMbY?t>}gQiZ?~kwMY%s zzNUI~$ieM`1zbrNR?pg7@+mDF&!($&Dmf8v$Cv`GxJYiHV2uEU>@lVqPs(9B=Znui z@BSP2oLtOoUXb^PkQR^s=NoWno0c&{Q9|YUm2qB8Pu`06DFPa4ug=5I`J2`GUwl2N z<>2Lvt*zus*jbu|N5(}K+rWxrn`cwbTcgNhUR?fnW}8^@0bz7d(ap0=t4TpxcZ-0_ zuszM#Z#50VRleV=zy!wU%Pp>e%Z|oRBq2drr!(VobqiB_fca;0x2rN{8h24!(dR91 zHs)7cQ?j~VwLsuHRn)oAq7uWtsH$q(9mrX0RhF$h9Nm78hGc_NKEyx*IG+ zoT2wWB?FjowtdE9BNT&A*hzhLgoQ2A`{i++D?ja&*>fg_*f(Lz-TK)8JTq)V|K9`m zj~nEM9g97FF~esFqMmj@l*G~u!iS5 z%TE4aZsfn1c#WGWA5WPr5J%SStE0@k&nw?Cg}F}yFzBIA$ji1jvF1|0A{6XCO6x3W zsBA+g?o~U{O1s+&X;>F(god?zt2_dm8LM}(aO)-Tu8M_#s7wf`P{JL%YND+ z1RRHrquv+!Zt(X`ihA+I(=M)LqX$9EWx%Q)t{4yxwBb^_r>^*w;v+rq|Iu{TQBii^ z+Xv}R=?>}c96~@kr8`8VyBR>brF#SgR8o+R0g;YT5RmS!p$8ad-g&;i_5RJ81^3+S zbN1Qix;|HB``HpNydb)96Z*bI;vVQem}!9dy2KY7(8o^h0((1I*L4#?-yx;n$LW9n zBo8DS#OzWme)UnQqt#B32O+}qF+rq=ZPTnMpE=tILBWX|qty9H1hFf)VnI#8IQ^OL%^|DUiZNRbmE)$=G?7oJjNY4XUZa&Kdp}*N z-FWbP#3p;MIl+LFn`X=wjLL1laK`0Kf(q{JZoh0scw$vrO+27%)?geE)O)?aG0C{?4wW1Ij8lYs~j+J^L}1i&MI@o zflf^HYwVx9T1F_8<3qRK5HFsn>tYd(xvbBtTkBKBVTsp=MlOF7C9n%tp{SdPDmi00 z&_Y6E&<({*vAy2t_s2Awvmx!g04{e8n(nmou6;o%jc#~}q$ls@KGvILBT;hEQ1R44 zVc+RG8TFPuqv%N#LBVy=s*_p1!JIVL7=)}Wn~_?L$QdH5Yw>4#NNmDZIVN$Bz9-ms zmxAF_r0Fft5~1?b{3L44f?qpJY^st4r`55fJm=Z-`XhMpft(!3&PSdr%A`xeJRpa2 z98^GhrFR zdUa_3=seTT#X(_LnsiGfy{6%z#S~R2y}VCP+QTFc^|L9I)UC^pd~Yxuo=2AXnhY)h zZV<6H;4bH1P%70m?6bx?&An4hlfh*qsFZDP{>rl^)L^sL??kLi4G}}ArBbyZL=fDX zZdlmIZqqRco`P3munm{_tL|_S69;RQBxAhVGa8XHt{N*e6iLd}75E>D9jBIE^>TZ6*hF)5e0H{`uG zY(09jX^fZpxhqhoDEKsRp^K`!RoK?P9{xK2_0{JjNB-}oJUnpfi9i-AU31Eh2MWiN zVy7M-NdT}{DQ4$%gz$+x0oMrdt261_{AS7liFw&_gnO$Zf3rkQ+TIJVo~_j;-soqV z&SdF76#1rCSo(vzCEd@Ma&E927A#wHy{}2dxjq8$ONT7AKWg*wyxN;ip6AL#9o-gD zjg`wh3-^-2Khyl1fqSECd-T^Tp|Jtz&PQT{F_g7VUfq zy1wXr(^_vA@yx>7G_f8g7;3kB7S-}&?W1*#U}EN{eqK&@p*Un*QOlbSX6GU26pvuhkj&K)V<&1{z+qZ@N=bFa9fFmh_Au1lF!}Bm#1Zv?^9qB8hndR;~9j8|62sq z|7T{nC?^?-Ql(u;ZyV)1;Qa)ix|5d5J_RBm)*_=~=gpJq=2EA>>nE^CKA3~<#nEPB z;km}y51O&xB6e4^u4;%KH%bhPCEyFEBSEZhuHhA>t+Y}e&-tfD zM%#7VMSGA-y0OS80{H5UfN}_m300~0d%Su8td;*c&UKeJwaYy1(#uis$K?QcyN|IR z?mQ(?euFtUEm(qUCo%YNX9XoWw=2PCrnW1!!G+cnF`K6B+n{fl%~xZ=NG3u)ePAW{ z^YjQwH~<(pEk0J4eBKC4aiqfk35=Ln1tOo$s@J5td5omJ@sa**)KoLdU)6&N#*yCR zjtHuof;%&Udnm|Nl5SNR!4x!r!ogit;CN#^KPZ!XUm3tgsq^xG-~BdSMLm?hJ^Lp_ zmwXj#Yj^t9p?3h@FU@Ah7XrV2dXEFimN9Md!H4iv?yLqtr-e$e!{yHOD-?fIn`Jtq z7x5~k!Ynt?XG~!$zicBhL}%^Nj*9NxQg@R6WA3}?iFpFg^4br zm96PQd1$A_9CeW{RpMz@@C%>=$VX$7i z`c@tb=JD9uoN9Qe z`C%MZjR|d|w+aY!cd4(8miszZ3Gu-@c&Ei}qb|1uTF{FKDh-x_Dz!?Dq~~Y9YmE!R z+(4e^6-;tc{IrORqEKTWCTrIlOvp4STdZd2pM#*%<&L}xb|;^8hR8ggfQzgImw3Kq zW{fa}PnJbojuBJypSGqxpx_1p|Is^#9in-@Rx1Uk(|0dz;_=It12?fcuPr>E(1>CB zxg~VN?ms`{?P;ZaA5r_ue}OUovSn{F5aWS_v>xocU+&*-3eX!gt8)bPAH z9cXxGO`S$6&_ph+Kh6yop&(7$)+3~LZrPvu;9izV^y04HHBr`710@*EvLI;UD=9mQ zdm)EnQKen&iq|T|wVMu-SX-S`s zexpg4Z6$s|J%h<@-iX`bVN7&8U20uC=H%It{KTfB;|LtH|>8uk&+y~ zF_<~|{W~3VZvW(v%+-GJC1EY;r2(9bccOTw1p*GJljEk9iDPsx-9}tgg(4j<3^f)g2|I!Z*st6}z) zkA~YSsPF_zVT#rl`lLYU1!PgGqZuEb$(E&hnmFp18Ajl|`I8~>ydPvYKhk&;!3w&U zwx*I`S5NyZ=2FaKSm>=0q@Xqacv&qUsrpQ)4tx2$wS9`2a;ziUNlIcF$s1QbIy1`<7tF}5?CGg@hwP+k=JDq%|+6bBsfUJm~gVca_MQNKUX zsu$?#23r_s+@sK!cSvbLt((hrFCP@|dxrlJjK_AQVa#s#C3$afb-$oVT|Mv(-A(Mc zj_IqL`!nR1TP@it>|*p9q?A)p^?Nu!76^$X=~lS?&tr$s6eK&8wq)un3fDc>K# z-@ijwJSG`fA|m1jgymYS6|bWqcZ~H~x=JrDa_0T*y?7(|=@tpEir3sxb{+xawP4X; zo!^c&@Wl`pRNGQ^q!{V?Yar_e`m~rzKLYB)UBd%oA=g!q6T5psD1FGEz0Z4ReUCj> zYxP*`q^rL!8sm1G)MJA=z9k&dZ-X5M6N+11y)7)u*{Od;lu+d&1sCKn&t}8!`ol3T}9nfPDJqSaeZrAE_&T@+tY&L_Yrfu$aYI zo{Hybvj!9V>kp~sy^sN08L)_!@@+P1CtU~6+(<#fDwhG{R~qzLgSFD!?kWHDDH8=1 zSrM8q`Gbvxd?!+%EKWwZIZbWhnM@7cK|+8F7=>1v_H+D*h+QFzA11NIyu=HdCsD3A zBG6?21gzMopxhSKkg=JeOOV}E#EDs6Q~Gr%IZSGDJp1*KWQg@i}X=# zi5T*u`25|S3X4!~bD_@G$yJ5FV0X}{v*_!KO)BZ_Gt71|hLa2$LQ_KyE|J+nf!H3q{Z7Bpt3%0A5u6A<~Me`>no0xaMZ{s zplt-JZLrkPmIa+VcDH%S*S7MdrrArmumWxAq$>38TrMr1AAUnx_wWL3fzHfOXa&wH z(Slv$+P9H!`b(1*vz`Tl4Gnt}0?+rU|9*Qt`;ZEFvh;iY!()*=vQS@AO7ORTb)7Ts zSWgM#UD}D|Y2{uI-ZWuI_9q4R#IP1?p!(fxx-FmhFz(yM-CmOp{|v%?Lmw}~SfvL;wHpM7 z!%0@$93Je;Zn&sOWV zcAZSWNd)|GNfht9S#-&~#BndKB;RO2>3)N%w%gMCJ1KFyd~1vw_@zanwgMxs{k}-= zW7y9Pctu&;D(frSPndDL_Lw)DXkOg;ZG*>UPnaq~hO~B1Bd9U2PNEa=C_8#+YbUGQHw?6=v0kNJ*xNcLqvDqG3B#mYqCbdc-zTEcs4;av$h?p5 z9B0^!$XxV)L(AMk5Lq7*@Ni9w4bE6GtIv{C7+$vqj@>Y*1SuM9QuDt%ny=?t5@_#4 ze=@ihX&=s@b%pEGb1PB_njeg(+L8TqI((*BivBFyzY69!H9$mTO3oK&^?VAn`!Lj0 z{VwlYhQCN%kB=L-8k(sl?$3(jt)Wy8-I%l8iM;kq^XZrGR-~7M&i&C8@_mLXZ4gir zXwkk{oI6ivnu@Yi(2Du7KgMRK48zf%M*ub~VQEJG?77&H|6`()7r^HdVZF zX2gID@Wmi&7qVLOafuLh0v*|w(!h#(O-`+MYiW(DP&NR!`|-vvbTIl-X;{0JjCt-2#CtQN9sskVn*`UK_aEN_n#ozknim4 zi51IgxcN(&@@v~d0y*<)HUylEeEuLa#V|h#_lLGUBd$wYpNNtjYhOBB2?iL596*U! zyijeQK9OQcK$x#y@qFG*JQ@fe+DZ&OI?|iv0o)REEUuv1kt;--4rP=-)+i;%8?q3sAXB7IGooF)Nn!s99oxor9N| z)!0m&|34og&C-Mol$h_z+R=%+&}PxtHP?-T{QMo@I6{7!b(>I%Kd_r=ZcfZ9&V?M| z*qgItStlu-dN#`5W7U8{1eYwEBLzzUa1Mm+4z+rLUHr{&=7P4Kc5|}4O%L#gC6nFc zas9uE~9Sfm&oj^uYt zm_lRMxJ~`{O2I1>(sBZbhMRw5fS%nWAoUidAu!3{^+H4Z&m&1R9!+wtHQVZtMK7$K z0knky^VoKXEE3EOBBxspXZJTF?i*QP;bG5~g5N|!E6C)R&Gz{Mb0*%X{+3lH$|BPA z{pBZhL%v|jVFMI1$-wi#O!>-jseU2Nli|U8SRaAGgpwVwj=Hfg zC~#_1wbT9I$Igc@^rO>1RU|wAtlj{i$3$cHrI$E+=OR#_z*HlM^;!94ZjdO66%|6Z zClKbe4ML;WkfH`|u>WJ0^+*Z|ZI_7gUx9V4{QSz<8gqLz<4$+S#FaMz*o{#mM4g=y*g`aTM*mMBUFcoyYm0o?F z)!)}enSny+7i3kpQc|J3nw@lS>Sk2_MlE!qWaI;p0#X(o_FHKn_HL$%Akp#cU;L^6 zdxCjeq_lU+KHJ&Sw8A5UzvuJ$xSM*wHMP6ow1rzsc!>%vaHX-A3vi9Y9939rIMscN zN=4gAaxlV7FCaY^HvSTKgWS?a@Tj@sEvV8n{=640LQnbHt@l~@aJ_v!=S&T6$zak4 zb~2v;{(i&E8Wj}OHZeXf>Q0|zSP1hbM9BVHAK1K`v7lX8_32IOFdc7tXx>qPJUyx4 zVXFQvrpZ0_H4?2%J_A`sF^n2Lx4;p8{GvUZLXoVVqAPkE`x1YwPM4e|hxo}KtICrY z9#0T8UmVokz2BEux2V2V5*5i@P~}KXf(?uQv3`9|@Z_JV#FE8j)*(tl@A`>wOf+@d znK2fx#V!7rS1k%FH^I>A20)|T{RZ1UG z$Wl88XgJ%)5vf|uCg7Q}8X^m&WK|}Z)au`5N|~KUzbJ&m>02$v8G<6j z&C%c47i4eg|CpFw#3=#O-!v#9O<_5}Vh(K=Fxdne4QXbkff4i?tfhrt=Ou*+wjl82 z{#`91Pi*pckA|-{YG(H4f*E?3pNEal63bJ1jmTMt<$&UV9p$^`+2&n>=DM!QLYO}9 z;hAdJijP4__l=**JI~dwZ_;-mz@1eOt=W(?&jl#Y?GB z340H}IdoUfIPs9xSobhys$JJ& zJCZYLJPVYJ$KwjtYA&`A2)5;>-5xQ1(d-qzJoW(AYJfXUMw20)t7Kz zcAQhZ&`lf+<8=rc3*0h~aCLh}x>+wA_<0`~HJ$|^M`g~>?ri}yU1_ZY{;aRoHuk-l zELi;#(|v1K&^g-$qT%6@|I1<<14@2$tCll%4mm6kz?#Q*!L>Da?`ilX+blb)@T&zWjKKLr(0OF!%IcjW2j8mN`~vb3P^TPLl-h*$V%DoVa)O|J zQ;Gy`%{F-I{4}wVE0-VqG*(g){3L{nFgJ8|)uyeqdHA&RE$c$U9rsV{9CdR%S_HUO zA-D?yNdup21iZ$mHb~8LzEJwo`pTI;Ka|I)>GNae$`bKHXH-YiR?}sJO?R^g-V+BQRafs;YQ2OoqOOd!d#8l6Z}v zM?LH8fEypvuk5_b786Uh33kYzSE=J3Zv#MJQ#iB@+9y=)ouBI)y-b5tCSD07rWgGi z!q`4w=B&>H>E1o`~FCeY2buRSph01`Prg=NMU5L-e-VHaRV+MjU2E$@cd05D95E88-5TeuFew=3gyy2 z1n=GhcEKS>F7xr+JiFjrWZU%>R~gEgrEO&I4nX>nElPKP$8(TLba5XRkuukz48XYcP2OZij0L&L};fSf}T$cOPpDo)=F3GJ2AuXyvi@ zYW)mj0qn%BTA_b_N}x~S*}*}&@+hA5AsxynZn8QLFrxNL8F-W%@+4jn&sEx5{k5BS zTexF0{JAh@oB?;9VLrU4&&@7PbTB1j6W`AKT~s0EPH2epK%dS(WtssX>CJ8QV3+J+-q2!k;+S$}UP9(6Qw;,_Q@4HR>L_&X}ATg zAzkGKglj4I(-9PuHMXF8jqU1Yn85gw_W5+~Y8bRWhbMwvRSO6q8os(Xix!wb@*tJb z!RDXg1t?V)$)@sKe~~OsxH^Q&D@T6tsag^ECFjyk7P%42A>=85&|U|&O~MDOm|O#W z`@>94+oml_w!%t%D<-^_hHqwBk8^g6rswgAx~R?B^KTjY6KTjbCd?Qv97bDaaC^32 zv-@NmLD!3+zWx<1r|bvaMXZ585k8LPJ~{OPN0`;p?!!jxoShSal=4#@4+Dva>XqLD z+l_B1+0MRSBo>Tb73{$EH|f8p=_3!W0dVM`&FaI=XIvSRwB~u6xvOt5osY3lW}K|V zn|=6|(Dkv>IIuY6ehTnGdeX46@@_s1K`%W|9G~>~{J*zMva)eMN5f$)-~G>q8w)}@ zvkx|8&0-g+VbOCecfr-b&Lel3y)T#s_p=f9?jiA^t5CMPAb;d?ZLf*Iu>C zB>=@1F{p_V2ue{!$W)g5bPelTs`eEkMWMbnx{t1lzMI_ocll?_yYJc!y?Qn@?i33o zB^mp?+2YULXug!r0CoXy)MZg+^ia?MFkrC(&fZAf=OVY|fzaggOJsfta&;GnscLk4 z>As{ak5l`@SECf`jppFBLB{YEw+1`#zT-pVP@SsHz})lgt@Vh($Htc{lIFY{>*YLj zSm#uHQN6oZdGy7z#NH*#`JYoS>(rLmGCivJv$Tu5&%=KQ${f6HN5+hFg|io!qd0BV znD+-cnT)x9~`e4h1J}TA&?u zfonNm2bTy#!x^AdrU2di#gB~PbuJ!-cX6OrKfu90fKcs`CjyWbg?RdK2{N3Fg zo-&(XwfZ(aGi37xNMA^Ji5dzEk~-t@1AEL(@`)2OoZT)yF| z0&1a<^$&1O5O8_xo|>Rz<_;wAfJm%*CH1tacsv&MWQ{~%thhIVPi@1uJr``K*fp3V zVA0tZFfzfsH%n{q1V8(_ck*tN9 z>gzj7k#`)2)D*m`y@4%b>=L`ERk7+y$OdRGPH9f%u?Tff1Y2gd#L$Bu7|7-bB64=!s2S@tTRg8~YEL!p6sNyT=D58HSj8o-+aci~Y8%1m2`LPNq=Y(-+l zavi-rqseuRuc2a}g02M{{S2$8gtVKMnEEOs?BjY+p{|h=gPC8lFi~gMNWB*t<~xTD zxy4dqysf@)W0?9jM}_f2>N6hX0Jdh^sN4`d67|E`kXI=9=;qD;x(P@kCZMa~Ji(aOMBA06?Ci$mUk}2WBKJ@v$56Lcf$_A`RhJ zGzTG~6*bvU)ybQ2EasP7yVZA`I@ikM?V9gy7Y9G(@x8vBXQ%7tO)>jsJpg;!$4XuK zsg_D>;y=>5u^;dCr4B$P2w$XR?o`{=^L!nI;=rZl3YM*kj4(Of%^L)u;MG&08~*Q# zp$bJYQR9N6S0HMlaBOo0JB(c%;}&k(=s!f;%EDjrsm31yP;Sp}Tj@HY=QQq9s4s0V1%QLUeKx3K(F7#15r~ZAW=BSRj z)aNU^+5;+UU`5C{kkr9u?dpf-e~R)4L`&5g63(!EdnLptHxURq^P-}?{t84QybiKh zZ{^%Boq1j*ysG&A3O`5x)AVH!{j*e!08rAy3b!(VTd}#hdGd4)MS}JMuUNfK*`=7- zU!S)E(kcT|KFncfhfuP}16$qXubu{RI6|ZPbq0o1bHj&4wC;kKI7UBEd73Bj9SS}m zSn)^{45^4nTfhI`pX{AF4G2fNc)gqP^_qn|pbqwILlEVI#-2%ddtJJ%2eB2;ouW?%o~!Y~NDrb}fqq3y6XE{0#P+je48^7=)qlQo#R2rSy7W~Hj^Tp;U4u)(AW{kF=)D2f5 ze;55*j^lbwIhQykC7$h8-H_A(lkz#=17464`g7&4y!JxoA6~p7P`J)EP$8&D$Va;e zx!&HcX@90q0k>y$_x7%qt)fU~;jTe%1o-dHOli2sS7P9LZYwL9CH>mQ=_KpQ!lc{V z>%TD#H%W{1f#}v9SJUe7dsmT7Jx0fgnT*Ylkjwi_2v9&Dk2JA@5%}awLrAv6GO~qF z^ML#Xvl^Vrxqc19@FM# z=U>aXJG!}Ab}vd+!5dk)-SKUP0tNg6e7$|1hR$ee0jK9owCJAl#u_&Mvh@3=RBaNz zYv0WBOo9b>QnSEZXP%Dq5B!_Ba+(_D3&9L=Q|1rrfdoV38$g}AO?vqU*xjMf(>_Sg zh*sgvkL#m3<1b%O>43$Wa3JETLViS)>Dqjd(-w$|va(vlA8AR{8l3OCz=O{u+ZCS zmK@xPrr#LL1BewAj3PMUTx>r*PGoOjpR`YMu&;>Rw>YxIpMtN77;Yb}t`@bjM&9_p z^AWA-?7R++g4#$~Se8-8&Z2o|2b10e9u>q;9b)?EcxOGdd*na+;U95~xvvqYXk!E; zPU$7O**q2wG2F`FAfF9}N5YXri;7QAM4XCGGA(~VKt&|$gw9lHVBf)R&w?EYiAY~W zX+w_*%aBhLu%+J5FhKA_W>QjC5;<9{lY$-qS(@xtQ`0N=Da|ihBP;Bx=jP>UBd&Kg z0cmnd`Cp;cqf<$z$ECGZLPfao_IeTUR{`dxeE`(3+-16AM%?;7M-NydoBYyci8Stt z)(9H=%*vlUiD#g6^Wu8Wp0@2 z%!ON}BOyPh_Q6W=?}@kgUUx*FLu`A)2CAIvYUq3{|Ek&^6O99H5)LCNlKee9WHM|% z;y=c9tIK446~*KoAYksx1Zr#K3r9To_c`#Z==t95uLBd2KSmZ)QE73Sx>(fx$ku1) zVU+|POPs@FN^PF^lIL!NSU(e8aBHMuJ(GIn(2NE}_4nFyJq?%gn|O_D1n3P?|7l?; zO(J-?R^fia1i4ei-yTvi?A%**n}eDA9Q!jWloK<7Tdci@>oowj=fIb7z3k%~v!H?Z zoex);Ay5<*2J1!2)xbs-=@2Be0gQSrhMVW27tFzMI4WFSTKd(+Jy@HOIOYLn0 zZD1bpo;7Pow#OuoCh6+zCnUN(MEJaPQ@JQdnj$V%+=a^$M{sY}eLUaGx90x15jkSnSb!gX-rxzPS7lM9nG-O&--3w7eBH zEC$FW8?&hiXtwEwc4hHWo8h?d}9Ij^EGgkOPyghf1^_g9cI|6hp!Wc2$~d8GW?HscS%n#qJb zaNW-}ACEE8)S=y2$0GIW;s;#Kv+B89PVZRD>hr)w_5!h3{AzhjFB+-r)xM|%v&%Vo z)!uDlP$(i8P0^I!>jMt~$N=K*71Npdad4MxM^Lc^W_8g>r?pEn8Y?Lj#{ru|nU!A}>5 znd<-qT-$*a5=nK7K$KqYnp^popt}WKYPZ%ZG zFxtuD!TiOIHhjDoX%}C*6PxuVzv@&su`laZx`h1NmJEx!ks&JbH_h+Rp-f43I)`jw1A+KLJm$&xzihQ%}-w+fjb9=C)dqxU*^(5oya)g^?`^?=!}lgY&PE~WSc+% z0kVKR%FnNvL0e0dCGS>G7syaEn3yRYFmb|grDR(ZdqoGddJ(7)BijoD;LFOP@~utz z>&qAwZ6rBfyeKwh>kdW~-7ZsM+J`M~#WRnFWuG(`TyyzHbTX6DYg?8eD{W+avv9)m3j0Hj;?MHPBZic4kapvE-FZR!{}( z?iS=*e<8ZLo_M~E)T#j#2p#HdmK$JlVDB+=ks#@POF}nJ)FB^taJN8I1Es<9fN5{z zHJ%{|)4S_U{M5kEeXxvknFJJ)NWwvia$su`6H-#au7k<;L2MR=Q!=p22_cF@rU!r9 zB9r}rBIJWJ4*p;+@j4j0ABf&JPy#_F;XLkN0jq!@WH>Anf_%Ek1fdlmA7|@ydk{!- zC;P5R!w4g>N@&A zj$Tpi{Vw9`;v1}bQp7vb))~FuXiK>K*q3&J^q=_&Q{X-04nCnL?aWAXV>(I(}kJVteoz~?+P$011`eF>!|xWR$pxtl_tpd z9-BDeA`rTveE9ckPMvk}+nxi<3dWR;y=J&FOgnc~HaGZpo47!r!&668W6 zZ+Z5Jhd=rkOm`XkEH=ArIN7)a!@cfN7_Fm92Z23YgFRAJDP#IWr1bCh@ zHj?AAckMQF^Fd=e8NsDj$byO~%erR%ZMPQn=$V|FI$K?ihvbPC3%%68JgRNf5`|@I zpYtc<+|iOP19|hKkf*_+IKsB!o@hbklK~qcm)$&iY8{K0s0a6Y1OZ+gsk!-tEw_iBq{wT#O_lRFAw)3|5KCPBi#pHqRR-m3>t(}}T*%)vDn z6(cs(d?DQfC7D3Bp9WcMYGw+GkJkm|SB*~?r)BpV{51!6YyU)?q;>VlbiVC^Zo3E9 zPVW-Ii0XU*X_25JnFvhGzs&1dBx3&x076>P3Hz(DTbXrSPI>ukTPEy2Q?ZfNe+2I$%K1?k!8zwd*l z-gd`g!5=eY`sH5)k6rE2H}kwjQ6E)D}KU*esl09X@= zne$O&8nwO2`-ve?y!$y7V(G^HgLj7flZV|`tNm}rw@B;FZgM*-v^%jI0Ob`|rE?t0 zRWs6$MMpr%&rK`llFjj3?!5q{;}qBj$J){4UKXF~%*UEKRa0<#FW~*?dH#YKO_nwQ z(|0UeqLxz7=p{L;xogudZ9BC0tL3!gDZ{>amPI_hrO1b8C#aR3|~FTa`8vn>amyzH$uxor|2D-@xqOG%5zMqj$5Iy2<% z#wKFoA5Ov8o&Js7o%pm{R0GOKk&@h(laZ|k;wTJ?{=;pgso(B2XUqPi8l3jNMDX4dKM|q z`AnwwiDW!Wnb+*==WAcKp|rePH0l~IFUAwVb31SFRU2F)m^j3pgy;}zDMD%w#0 zr<$wtlxi^Y0Vb!~5M=#rROuOF7P~;4>{qMVf?bZ+tGL4iYtgruw+tcme=8wB@I`)L z$bkAba*X46(6#*cz$DE7E-t*e2W!CEvPe29Pqy?!Ds9)QqJ5#QldRsy((|>6>!(v6 z%9T+WBb*;b@9!S@vflfT$-8M<&&=CTh%LC-AR`equ)BMJTYjac7!rO{AodrDfP%rJ zJ!8F7J;D)Bmyy7w)}IZ`$lDt(%)%Q?7E_V5Ga2ax&9}2zIhNeZp&0~ZA1zklgKFn)Et@j9KmD(W&Ds_9OU!4n`bmhVS!h{G7klNH=9zqr+H_=aoy5lbo*tdI z;*y))7X+UR-wjOf1FH3AZKjvE+79*pQR1#fz5ji`ovE1Z3Igg~?m{vde}kNYPfSs{ zQ~7?gx&o_+^MsH2at;d(HheiRbhA52E2m=AUPhqi5UXA3JqcyQyA%02M|194Z>vcc zn;hOd6Ha}ZC;!nF68u}KB5xE`>^y)7B*y%EmQdnnC#@EojkLIlf=_QX0?m8joPYlOxkT+jTBraj4q}y_yEXdQN6F4}#N&_il6H?gowr=^ zhecwp^CI}@ig%}SELZ!G0(0C^FZWqfNzX+NRO6EOp^457<<+)~r{pox!P;av%gVAJ z!|Vwk0gmVL@P4uDi|FhDczoH3uZ*D-b=%^&-yC7)5Us=IhxZQ)*#(CU-ZlHpy$i(o+im8(&(hrw|1>5YZq%yveCJ8|LGiIPjTxZV z;QmL&QPLx~M55+F|gkKWy>4F*wWb5oBs=8d|PhH}?&B)iYkqtKOS|jFke8G+3@0gaMj7Lh* zVz{&#)or~1w=9BRIVUgDuJl$~A8)12e?95d-%37F7(e#4fw&T?sNI9o= zOe$w8^?;*8iVni9`A7G5P$!~`iX4B1Q8D0ULbO>MntnRBt@W_deQPkp)(tDn8T87y zN-N(K0k&7pMSP6H&KI|HB0gJbRlWDNf=(&D`oL{I6A49MB4glPaO^|kGL5zsU;$Cd zxFfkp<7DNDTxz=$E-(^7Pbt}1^=_Bkb!K_@okdx1XcAhvw&dI zwrRch>xDoJY(lj3x&jmmNTaeqCmWZfADx?4-V;K9)@HvA^cM;!5^=R4;?aH3_X$8wv^TC`)t-ZgFi#5=ra zhkvw$T{B887re}=BW1l9C||8RDB4HGJLjUH02T9bw9&`odAyp!dw2`Hyp6+YOafIC zxmcM<6FPB^F9A>Ua^B;W@u`8U720)PHr+9hmrm)$wZ~`+#>9#U&rgZljA1is!U#!LUn|2`vP%#2WR}n!pggWeA?1PFH zgf^vqT_5Tu2R+$*#{-L?f>O(S{+?H*b|H`p3xEl;|GH-&@M*9MN3?031My#fZ3DkK zhxY-daO$p(Q$TJX7r;0#iVGf!#h>nU^iuZQ+pIub=@K@-*4N7K%+~kKa6iBnITby%{<;Ty~!$0&=Ix8=8q?=kSaal z5#Y8h+=5g}ahM|vuu3Rj_DSt1jFl@8o<+B*WYO2iZSq7{u|%NeH%8n*Wr@L6+!zK` zdfbyn@Fw}-Vl8?u91w&s(OX!3FW+h*(^__ zOrJ;PQXyW!Qh4?SpK7vE1%mUtp|gqji%m_2Nu0sOyp`41))zIugWs&3+tYF;7Qf03 z{#g?|yOs1@{sIl;i_3;6+?Dbt__;WkI7IIUoJ~ zN}4P1A%W=lQ<_-AG>P`D>If}0e!Z6xh=f6!#eKfce*_?vOT03${#?ZZ|M-hZrQfH8 z?M&6TwkC=oY2{1R5IX#U>zv+jkExVu}hy+7{n ztd%tA0GMh}n*O&9z;`miVs~Ks!R^OK7Btb^CkahF}-6t)qy)LAV&Yyyg zzX*n1PET?8kEYX%Pky6>KRwiG2OPvWtoC?z{w+7vCz23q!oN>qQm0DDCRSc%aZp#w z>2_0z7%HJxcFgyeMwLPw4M$Dc|DEj`Ilv?r-zOEZgwH@nSK@nvSAmj^2FvF71L((n z*!*QNHQTja_A)$}U*x6T$zF8bz8ZQ%$}0O5IdB-bGAC1L^NdDFqv28+6F>5Uyxb%H zy;OFp=dVVhdCONJ7A#p(#KZb*s}kNS zG3Iq~6#wn%DmvL%ha8taxi}J90V`y&lMnO=zr>{q>z!Zu3!4XKG86tunfSKI)mse* zXBd6LbO~0FZkj5v0CHRpv}m6^+2k|PICR6pVO*q}3AUUqIVUbT|ARq(w#7^3zLH`| zVNlj1ES}DK&#;+JHbm^#8PZJE?C@go!w9_xSVnMrE~ywI6}oZG5}w36m%)l|`0io4 z;ip!;l`0)foLVv8%u1FUHm2K5<}4a~Xk%{i3T34Iv9BOpV&zJlihk82VvrZQ)?=rd zeey;H2OfAP>-CJPGHO&Zz2_@4H{a1JlpG#UqL1==8;6uh(7kgQc95VIl1y%!ewfM; zLcb4^d6r!KqAN*!CK>j#FlX|6qm;vN3Z3_CEoW|LU&q(3Ji-AMN(@@h_-l$7NVzt6 z_R;5Z-SBAgZ#hsRiBI|DK|hW#ZJr_p8cILD6BZTRXnRYJ@M24IUi;&D@P-L8p2fjd z1xEyLmAF3;aE4v@nGkU@gLC0VEJo-adKZ-t%0>)$V6GCda}pO!sOf$*Ita0yNKb0r z#;P8*PEhtNu7~g^Qne1bEcvZpKKe{R^lL-Tgo;y^ScB!Y#sTn$YKR7MZe|UtQQJsK zA$#hTOw)7tnaKOSB}!@OqV0DQ&!X6(zYD=>9L7ds&mert(p-%XvkwbUPw()arKDsA zy5^cb28AP*YD*U<0C{SQCwJ7X71o1Z1asQshqUW!w{0qN)Cr0DXjZP>$F^z=@~fCZt@X9@gIX)NWj{b?+XRJh5xPHf|)4t zEk~h}ll2wd#bux`K}&TqDY|l4R(p0;nPOzFjk4@!hdEed{f=Y~-VS>@0KZach@EWc zDh3Er@?KLEWK9chXA9Eh;rI$aU196loq9M4%!J5CxYhRnrwM~-pl;Em(8*A%*X|y~ zW0rcE46`cj_M9^8IrM|s@68!_@!n&WY6_@8SI)Z1hk^8;_WPedA+yO? zsM(?h*lif~V&|VFBmnUu)Pcm%q51RFmiil~)$DWhxFE^a=KA`uCO0iNgb%vI?5@{2 z%mtcnB1Yg6)XcK66Ut@=e9TTcU=AVF0gw8xQ)&&682Eh1^V`f31pFfZd%G6;mG7QQ z`JVf|CW&LUKQ|h=Z&_69&aw>WKw8!z@=U#dqT~ZwGf-8%eZOrw&Jrh(#REP*4}k6o>FD@~fjWo~wkAcj#m#fc*9WPS9& zc1ky~w&{jKcCyJx9&Xo(?^EfP7ux=DWyh>pN!R;oVB?P=&9!1AcP;%iYM~o5um_G6 zR65&f8HUsV-|4{8gciA-7<8sAtr1;@+NbHsOd%H!A|c`sIhYoqMTx3wfLTY=M3u6% zg0om~CM0B7|CvuNbwU>rTfoJO@JVI}yyd%~%G!30jUFb*j7_paU7$SZf|A>oEsl>6 zZ#$718k_p(5OH&G2}!CwXJv6qt;5St#(gZqu+erfYCr4~jvnXw50x``J^O2!vfmye z;BPxVQ6l++o(Edyk3{*ajrJK`4ggw|dlaY|;=RK-X99Z{2#>HVL^H&O($7aIJ1+Ru zfAa9pEo}IPH7!ap_ZA9X{w#0&N5CI#i5`3P(AhLKI16o*<$@sTVR3RrP`^g%M;yq| zvf`V^#p<2=GqjmGfcgw*!+vG<;k+|-V*V5Ke4u|mV$SY;wlthMRN0IKI0Qsh5|e+0 zG}`?61vd~wIHAnQOjp)a=&i6|_eTe00?BgH+5XpPukPr5R$VbwppiLEcDx%GBE_?t zf8DtG56(P6Dq=_b%|n% zTrz!JAS>cqdMd@LC!LvAg|wrDWqQ=_qKF<6_$`yKnbu7RmLQ6Bz`2P4Q`M);+8B*N5ic zO?5-fj?0xf^3c5?@7OdmSboT!NtY*M*q7~BW+>IY@|#uV!MI?kIa)#J18nW1eEWqk z*g=e7ptw1OBtfGsnT;K4Lb%Mo&=FU=!gYB)6Feo-r~cCnudUD{_+dY6zJaSM;`lLI ziIcTHW@j3#?f9rs8f7~$^gYIY7;_{;DLZ8dlY8)n&Af;GMOOnWm|8P2)g#nicsPWZ zKSFnuyS@BC70J3bC;!riG*y!3R)|WV0tGG;nVC1@FU3J=i!h?VIpW6-sE*zF_@6}csqd3 zWktD$P$HL{SN#dg3H!$NuwkcH_;+yT&&yrzykc9{$+EgOM5X*Ii(E~_CG+1H&-uX_ z_DLB#oS{eW>=9Bi+|XH7SUon}RwtTQqx35fgD74HFMWV|^gEiSQKk(NrnX^jbh{r@ zDpS({3r57agMsfXo*lj~K*&Hx)6#0tpPBn?xfcnFfLd0!!=x)i0maCEqeFQEIK8DG zNX2__1omkmv12;{enTK%gLCrT%V@c((->S`3a z)+-8ly#j$Pf#)FepYN?uM9L=CS+owEsbjK02@wlW-iTDg_{~N6ZY~6}pSgIdZ6^nh zJ|uXR!U>yz5JwP#fGosyD1U!OBFpB5NglmMeJ{3~1U|nP^_xNrYyU%_Y0>k+(NW6c z5R?aooUn0KCVmavfx9!_W0NBl)!K)AzMtD}2!Y%;kolMY0S>Qc`MqGFE{&hNB}Wi3 z@C+d&5&j!7yEcY?XI4Avp{GIZ;YQlS>sk}ZygUCiyVO6gyJIf}!^Tp2wId3IYi*if z^!F+3eeW|%QqgOD1h96ZQxN=0^3<3((&3=0`D%28<=xuT)R(QF4I%b$X!~o?l+!mF zR#lM&C4eh2x9@*tInY>jYoevifU~*h>~4Sf^!K`A#Sh$3ew~aQ)QDW^!CoG&kf7{f zRb*3jeBOv{@>bw6+B@?jWrt3-t^n&Z;wj5Au`^Ui{0M<JKIGdoz%hs*rSq6AIze{eVMG(v_<-wA>_ z!?M#BE2k{4^8;VnFD0P-rjFUE%*!x6NrN`;Gam}k92kIsUsFw486vYe(xok4U6DU$ z!0mF>wE%&)Yq|mR@8d&z1{q@k-GN5&t%T4tw65uN+Dgp zp_r4J$Qm>Mhf_pZ4O*0qt3DS9SFw37WIPZDWw~5izNV*x3M=@ssI%#+C(62;Fhv9& z!R>)oM_EIbki*=HziO0rlK$XhTX`YF0G2_B26HA1iB8a*RG3nuTMm~SAG@pp(}u2l z)KgHL5aJbJ-92Oep~e)|t+UCz+O%x{SK9PZo!qb6I4RK%PXFVbsDs=PSM-^}!nA(0 zdoDHt6epU1;#Y=UJLD=@-0#09&MH6<^L)%IOK1 zug2x6f9SL-lJnPO{DyXOT0d%9eUERcP*mF2f{0BfH4}qS(8kaXCdbKLyUIm+_6m!S z!*h?koMmWvsyUR>_umvWW66xj>Nw7pYxMnhGhGaN)v7a?uQ^QGNF26pbxGz~BjId! zV?`ZD|NLF1t>y0P>w z{V_{)0RM|OHSCH3w4PlwVr zGK%{@7WiGlUTYS&i_Vs+qoj(9jC3^06`mr()zOiST%02hX7{O?*xq$RwESkLB{h_; zWXO%eu&DHv6Xkx84aBfYWndXC48bCq9&0A5=>xn}eaH!1eclb>*%G;W-=qlrb)-gN zw{$EvQ8NHvWSfFGQvExdoX?i47rldHWQ{W2FLA{-kdq;r;|luGCgu*LUL+G4N;pfN zZ-A>FHP%$ND6lxEZh>H>UbApj9XJ_E>DO_WpZvj~oL=#O{7jTqC5nA8W5|;x9|)IW zuH!Z*TxTV_mlu=EZiEpN{?S_3QvIysdkUih5%b^}&+XwKWG9X)uBR=oG;VQWDIrV` z-HwTjS{mPxg!v{%>jUevS%G=8^(BtXNzBF5y)Uz#qiU2Z6yY@9+CokheSpD~3rpn7%?cxC$hkYB^IovAAWi;N} zMN88cItZ$WZ6dcw8&*ZE?f-11wjD-Jiie*NEG~gtmKvHJMpEz>;iVB8_)gz(D$fGulGI=u- zRoixdgKY$(dh_KlN*y>P+c6_fr=P@*2m3U&p14$NE{=Q>>N?!-w(%m^a}$^R;e81> zxpxlQbwMHnyiQLFfHv>N6P2>A_ie@3GtP;ruD609NM>1an^nJas{6D{r7-`J2agh; zz-~3ABnE=*-TK{cpYsc4OVEG3yd^j&0Ki8;AankGZ?A-@{X6*Ys)M-s!gq)h-}HPj zelB{5VSEl7Ud~e_c*C2Ga)~;iRK$F*Y$_tuz`_xIXeC~U)*JUp_n*z{*UQaHmQQ4z z(js2RG#6{F8PLn5qZP~-nnSocT;uvcpG(J-6PGW@;mbSicM5(Kzo$KMvHG{ZrA?&$ ztrbp#Ymz9?NE7)^%xQukz;Zw-Qq;|yRoPEW6 zqR+!?N3%hP#hX?K4vxR$>vtm@hNKk+Q<6^Wj9LShlAX;W8lS!S66C-pZC_Q$DckmZ z7#tI;BJ;Xf_#Jl3*G<-Yz*dQqK6|v?N_*oQ6jauP{cG7an{{)f@_9D>GVxeHZ@>!e z)YC#=G-8>l&M9ANC19nn6a8w!YLs&Z5oOfHr=zvd6?hrUqWzoK$InyW6Aj@on-CyUmh8ilzzrot+Mh((kpaP_u?HB&Usof&oIY^i2}-{TE6IMhE6l8mXu zh9=|KF7d=ppu%a=q;X@`>ErB8_;^%hDEN9{=)4@XO-JU1gKAWiY{x@@NYHQ?%AgLN z?daC`-Dw~1o|~aO;NfErAT9>5z?>DnUk z;dn-OKEbO<{oUQg@%c-Ps5&C$HkNxhtSM%37|9!lgAsPpf9kd(|zp0L%nG8N(T$3!aMO$OQqQG_A*h>i;j=e3R2wKM9`Oe8rak5asZ;z zk{gAKWi>ImWo6G=Iro~yBrk1dfZv{+*Ix_RS1vZ|-07K#C!E&OVF3hfT0dn<$Ow7( zDwbh3h9|$wT7BGp`#``VN*rYmfAA+IyCBV_Ok-Ox8BCJSXWO0rlY93K_K#G^r zWR4YyI_m|N6}!`7MM(g(#mhr3B&b5_Pzo)9p8+w)@w>{HHa!*)s^OsU)6r>-iVOQA zR(W}Z97>39FQXUI6zUQn-YHpxi1SY1ifB+BI?OrgVQl|X3u z2J2Uu$Z{mNO=~7eGO*163JK@^nWxHpwYoypDPVFn3hNP%Xefi7*yleIB8RK9!QZCi z^^LnQyL*PF`)Id*p>+^%^NadL1#mk<~pbMJDeX@E=B?AX=ydwVzW|s zgc9oq&xaGm@6(%@3o`B_@5j1p7!6BmsyW3}N2hFQSxO?BLGk^0_9ILVgDHCjG!wV# zk++RnCxoLVrCItT6SM6=%8|qgncaMIIng3(E+5UMkrnOL(T=-cVl};@(VuVnf(%3b zG}}FNSv0#cZ&4njs`oNbX4dFFBZTFru^Cgne@Yh+k>uHT7d?yBWBw|@!E9YmrSWsR zoiyZGc2bg6Q7H`B{p)yX`T7)Mgcmpa^{KfwcHQ}mGUB;OG^N#ec%QyJjonuJ(_dM& z?5~YKJD&`C#~PJYiR&t5?KCo|?uVSJ790BgXtzLzTcBg7Z^{ zKjs~8v7;7anTkc^lH*8MF@iY@$rNZE(I$2`FI8=sIlS!dkj1HQhkgJ({{K|p{FPl_ zc1*cwZ2wm4f-CctGleLaEZWc0NfcT;y#H_QFh;=LyG^@G+|p}eKYWp{j@0xJr1<+u ztIHk8MaML`rcXqLmqau~mXlS6{qbG1u=w$3JBe1tK#)S}{r~9US6WlOoTkEiZ-eT# zyf(OA-hT``fLpLtlcJ3t;=fCOZhYR^1xy6AjUwPerQ|4#}*Z>4tRU1*Y+@ie+~`oi2r=^lX)5 z6yngKI0drSwIKZ&N$^)^$OHUC4O~LYoWs8!SWVhid^7eUSiki;U%?GsN4DYJB@^XV z`Rn=6rHl>A=!BGpLI|p)j-`}lID)swck}I|?;}3hT!BBcrhZ+jD!;VzLAnmT2C{;t z-M&Tc_C`SNkHsUFl{47!sXC}gaV?Ku2=uwY`N!|!9gtl$e^w$A(M*9G-h1_@kJ3?? z+6;#DC`a(Hq%)Razn!~L5otUVj6eC}*dmjK=b^h(^ZB7_@LRlS2U;bK>F7~c)ss4B z?Y~2%NnmVARuJdJKND6PHRGkjJ0CQvi!W^K+lQ)G)=npgVuZLl8!(FgzAOTxzG?43 zRIXXV-6O~QZ=Rp>^@w!_G;d@Xjt}-B8AXu3Oma@909I*iYOxBJKF;NrZ$hs#lAzkb zgzw)Yr0PEq+z&Glxnt%Dx%|qjJj&gvXTHG(Z2>7EK@g5--^Uk!hbd`4JHy+qSBMNy zw6fa9Msa5(lyHUHvXr1cSyJ`$l-&`e%m;JR<{0{ZJf!ZHjf--!ti<@}Y4n$ni(|aoGb6i+uHeJR z5KoGaB8ETbFsI+t(++#Q=)o`aFYnUIF09gHz(6oD4a~V;W&N8FVL)qU zj|w#i7)Ma&YXs{T5_$@9L3Sbi)W}FAi;G+M5UdTd9T9{n&OJxDvMokw%ph{r&@W7BR^w&R+(&uaNCXYk>(Pq^P(#GtO)H*{rkq7saz!@wvB=&y zt7IPAU4^{@G=8gPtMT`ZS9X^uO0!1@=fVE z3EyAO5_-WeTh#k1Vh|x>@E-fO4zN(p8h7Bk62GQNE-xTSloE0zyj=t1dyP;%;I{;l zeF2wJrsJ;og>-v{^*PHkmP{D!O z=d=hKOi}Y*x*HyY1AP0=gRT4-8=OLTM0!o*pXXGtg8{x+553)BKxoduNJ4ZpXoto31gVNTPPxU1go&zFt&mAx{m zirqm8U>J2{o^B+d=ah9m<&{L~H0TC67uHPN7pYphf-r%o+{bzOX1@3khhk>*(x%Y&U>a_BeUWm_x zK(n=-(eGig$vB7KnBLxzv0ye2LT2Kr#5`n7332V^`GP<>uRsKQi)$Qj5rApjw~_p` z+#irR-Q|#XWDBaA%1NO?al-c-96*PVweIW56^P*4FAx5Xo8(JZ0Ej8zc1tGb&(1Ek zV@4==dI`9jaKuYWVQ7OmjypT9nWGt&&(Zhsn3#0Mr}f>aFHralj^?K zy)6XSQMaSje?&@uG6&2n^wRJ{t$5AYF&C%f+JkP~ZOE1y#0Sg#%Zt7i!%^GR9Yh>^ z%h2CjRFG(!=>=Q5sJ5(bHp8xpSMAs6in{te&?9Ff zi}+fFQ}zFf$4Iyr6q)8_bAdUfng0AMSNVBRonM440uu|w$@lpnLsbTSdpC7azGp7( za-nzkb>^|(3!{yDOs{w)sPrO3az*^A_f{M^3N6+h2!`LtNCdw>jG6;E&TfzLXMocq zz;=yeyr-GxUQ-)5HbyPZofm0P1;!lGF>$9UQf_CXHj8dWYx6X^X_pE^NR{L5O_*Ho z7v8A5c?A?qv4xtUsf_9CryZP?qkJG0G6(g7o}f9i;Q33N=;Dzg=|_)n#nxr7a3JqV zX4lP2_U(shxVaD}jC?2+gDXohgJEyF2(a9{zFLw&6Y1Q{v`U`v|^op5W2wdbhXxg9aKDC*u?_ z1BP|o&nwTL6h_(HBr|N1sT*?Pci#hWkcC>mcvrsNPR?ZIi#_vUhcM&bj`LpHOrWkt zz}F`1&05{Xrd^5lFF>jV-P)g<|F};N$f+aYKMVlO{h>*862Oyx-Dg|d#xITUfUE;O z_nv!k+lecW4e#aC{#C~_;f=R!-V^lp*P71hf52AFyr9*l#)tf8HDst{N_{_m&9(RI zb??h58>x_0xYzltScT<=9CaV?an8x|+AibW|h7N0A!qAZXtK#i}y z4D6v-VuFCe)&plYG{zV|jfXe(wXSx7jx48^L)D143m)=5Q)kdhM>@1LT|-q^Ts(Ga z%YYME;Oo;>*sB;TE$;lRq7jHOp+DoG7v7A#VEpm`?uDB9Y=qZ?5mb66Zu~R>0UyOx zYg2#HUWNL$vOOOhm6 z{niNC56PuE{g5BZx$cJaX52H;W%h!e>TOLXL#rhvw}|k0B3C9wD~UafRJKSNKF~o{PIej8VW)`J_iBc=wEuJuBk1^ zqYc5<_^Non>lEuMum>TaG}Vhpp23I9D&%PL$;7`0f{-|6q|}gsTw7JFrGDf_?HN^G z#lu`PRTpZpq@DkcXrLfOLhjHy6c<4Rg0m=XB?Os5H;7UZSfcgqnS}_FP!OgP$Su8QEgYjvwvUTjv^V+zr3D8dXv=IbnflBAKP zox?qU+r%vNyS=eNQB+rw8gNNnEe++i?H^S~xnn$6>pRq``>O9wr$5{QeX*zNl|rEyaa42E~qky_ag_d%Kl#rU`s=-@bub_S?L+T&&0j zn%F*xBKRHE$`$dET4^D6 zgUj3-5=TJ~*(z1RMf1aE^k@~so)=?$om~zG`K}^cmi}g7D0IGAp$s&*J)C4qW7SLk zP>k!kevprmgG)VH0Q0Z=TkB((9GFZ#viC)F zRUWOV`VGbw-SG!*g4o4sr+i~W?#)f-+$Q|Lgs_#Ls-nY1s!-up3N%@f1@+G$?6&nt zoF>WlzS7+oYEvw*wKacvU8e8;=9TaL+~uvFL>rD6N93oSuj$aV7su{(!XD2fP9Zmo z-q}wydDfV*8AbB5=1YheFbs%;3`pO5h0HjqS4Zr zkyU&X?hIzwbTgcPM=9Ssxew)a3p41xN8u3$(YZm%+PW+nbvO=#ed1oAMCI653F7q~ zm$TM;^?#E&LXGDGz6NY?m$OVYSi5RFqHEdVrq{)a^Pz=TzEIXg!jfb8F$z}bW|WW1 zRu5&HP{Cfa^(a;o6XL~ZI@E7?UNweqR~wD$tX66^GkKJ4s9IUHW~R0bWR>3V)JziE z6cFoz(e5^;5Js#bVCP5TRRJ}WT_Dp)sGvG|6P1bJvXJ^J@GxK0uOXF1Th1(!=A|@0 zD0$v?Vj|nM3sso^F8f~njp>(ou_3ubAg}X_4vTh6T-U>)aWaRK9y2+>LKKT7VA^!L zUiu5F?Th4F(g+rnAoFOlA=l{;k90HHqBljGKwRJOtH8B2hurx(ty)4@spfM7wP)hp%ru0)Kzp9)h}SX^?9gA257^BQ;Z@np^kO$0 zjq(E&QSTdvD~d^(0e6QK2L;dx;Ns`xr^E-Q(>d6DRh$nV+C7fu!-fQ-e8QNLz$@mv z->*^f_e#n}8Q`q7yA0eN{}$t(21)tfAJf`Srl%?CvQ8442ZY=fAm8UJB|tL+v`?4H z=@%<>QmkZS#}=1OFsaN8j9OhH8l4;9Ml>(1;e4x5B|YE*nZ$%L6l;eN*9L?J>XqQA ze66mabghY+^=hYmZTbO)J>1R2^JsIdXU4T6 zq$>78rC%_~XRyIGqh_PSDL?fRCLa*FN1~pqT_Cxe(e=UdiK_W}4r5!~#eylPpxpPd ztGU)qfE4~}`t5y3FVqTZCy|E8TEBQ&sf@g}5Ii{#r0!8h7XzE>m_YAZYZ$xcglX|@ znE9%YQOTrae}aX+1_c<$;#_jjLmC>_M7GkP<|`m;sy(={<}9*@MDIW=q{D^S@&Tws znEFG%McMFdI;mj#i-ye(AASiuLKp-Axj0z7W_o*NHzQzEA2s;&9D^xjKSb{w&)Id`*ui0J^y&piJltM3Dgea#RZ}hDT^X zijt(s1HWxSz#`OOCnBW5b|Fs4?gKLQv8cNc&aA-1sA)S)@4si9V`myJ>#WreNebW5 zDyx@uQjTXisY&>Re4A^3QVn$^7B~`_2@j}?dTCf$2~6-koJ|6@0&7F1(E55oR>^J| zXOR!&B{t*9pDcF~ehL6;x}e#OpR7P`GRAMJa873zC;D#x$K;n~6Z%U1 zTHMQ*((CJ))kYibuW{?1Gu^BPYQZq}2Qg{4$3+{qgONWP{Qt17!oX(q!?_cFP%g?C zPv^)kDfHBMXlZlS>Xi&M?6vHk$`(+^l1^;I<90Zi?F9ncTb`R6EOy{dURBI@e$w{7 zKX#&p#S(EadY>*yzVLWJQ#Yg(L{!Q6zOcvsB0A-~iNqv_W;{Nm0g!s=dF-c?KjXm7 zCGQ~`{32k}NtcJZo-56xN3jYO5 z=V;=x&&{#zM-ZFCe76bKU{~Cf>M?2en1sKe)(P>*)~mMN8DK_Xd{U@fYMmI6MYlbf zKbYT}qq@E05?6#n&|2nqiUbR2f%-{9b=z9}|9mm*uqihIZg^*@F)Ts13M`k=%fne| zrw1t3gYcih)v5mqA~geG)4LSo7*c^TVhNB0)hUJIxa=O`>$**L*!=-(M}#ax^Ky8r zwj^$P-Nxy*ul1XMXFeRU6V_B4G;0Yrxz>hC4y<5@Y(x26AAsODsWTjc)ckt-0ArIJ zs6beW>g*=7>*WiuozPkbAMS!W2&BoO@LMgH@s;LprD-S8^slj^qhO8x`#E{~4K7WkbDT(kdM0^zZAk2y9@RcPLLLLsUF+^;vL-=W2@JL;Z zwL4teBPvppMR^J0oc)=#Yk;pAA@InGqmPsypu%c2d_pRRdasj)^U%kw>E^b!>QDO{ zv<1T)xBb`H2hC7LUzgwA#P3X!_GWK#=D5uiVl!V!2SgbL`9i4(8_JQedkD=)2NXPx zxpLndotP1SXFp&~&G?heul(?3(H7OG`RVMbf!(M};{>=^!T0GnH>82o3o^J*O*_Ub z<%}j#{)UEyIrr#2TmzcLsbh+QoQL)`V^>v}4U=F~e$vU}bFLWpf+i>cyF>}t49PWG z&{34?m*Rrv!@_!%r#Z@N+gx_F;}JG1EY+B?{U8%a{d6!fK{iX~$c8=QfcesBcR|K1 zuQVA&9Z*J1W7rT;e(gQ?OI8UUsvqg|JxFP*Hnv%uSkw@eQ5e1@v=wBRiC*T|X83H8 zG*>#Kjh9DXn#yXbly~vU{x69R8v8B>2%S43C^xE(QOOsQwi0kbNbuFUJz^`kD1l28 z{VsJ>U3TV}iitf_e%ELATf(v(eWHJ5MYSUvGD^*K)gNw?4I4iowUx{lL0|ZJ%Fx5w zo@b>Kt}p#?26^|%mRHze-0yWZ6FG|Mg>R-Hec~?WIN!lcEQFcDZ84`7cCC6XS--Pt zK(X0||F-*?%?k3>qzSXSURAUnc#_AMT~$l+X6N;4&!kvIgXBWXsUCMwemf_f`=C#U$ry zkvU;4=O*JalvqW?+^${qj=w)`FfF)C*#G#l816Sd6Z*+Uqw9{M+^Oy{70FaK^Avk=YqJ^8K@9kSeZ0fkRGaX(yMogfq(lmOlvpKcl>AjkMFE9! zCPgH@S$j1rMc`xbHxK z@8FLE5eq<|;hT94`1LvjmRp<}VmB^u4-;mAZ7B&D;KLoY@MK)bvzcRCoS;jDy7-*T_j?ajF=RZxE+AeGv{q*K z)c^3f{7PG7@;tU%+x3uOoT6gwC(pL3f#fVKj=-B4U;Lxv=gq)=)3YkeQ#a$C6^`Y? z1vV_$UxUZPrk<}@;B7dT$Qt0*l{(Dnh9=h{b=`?2H3#JlC@;%*D4(@dGg$wQ^SJu_ zbUgh?f?|G4$P6Rt>W!j0c`W(G^Kd3jwEgy2?Qz<)LxBUOgQcMKTQ%bk@=0`@4x7nz zOr_KW8ki|Mp}|30a-VVSwty#7fO6;Br37S_r!?i;F1{BNw1sO-$rAZ~!sxWt;{2sB ziL|qmM|zBefZ{`;ph2hSGUxnlWIB#iZKRboG~Y#c@nTsXXJ9Ur2H9ujDUv%?sxJV? zPW=9QP(zMAwE^eFqm}Cw!nY2$0qG08z1LP(tfAPkE6CZ&#Vo-_oK0iXk4(G^j8nu3 zFR;~xc@{*tkgr33;zr0%N=niq85ljl(Q*fHmNSST1U{_?CQOCGcoU{d=xQKGG?)YL zeh{$h;D}_$O0MXWc6MP~>AQpm#NJCiqbrqbm1xgV&wK@e_7VsIlsl}kI~O6YXWb|S zu;hgXP(4D=RNh_2rHj2WKms&Fx!klpXL@q1F1KAN_lHcYU5Su|N)zt7Crt`5R)G`+ zmjisZ*sc^q%mr&{=0nP7m+IwL9Q(IMSGii2{XOvgLI2C}4bHy-F(xX7RA}aq;$aoe zy(CD1Es5XRXwtmVzS33#%J;QhUvcj!5*9y>(i0wEnNyj7JZ4|uLD$$1)mV!(J7Ec<3nV&pNRSI?Zl)i3B-@&&5aUoY2 z>DfjtsVy1~JqaJH5EzisNPm#?X-Yby&Y~WD1nhfmdVm*R3QQtAv=F^yG7JcYKBXd2 zvtX7eTJM)$XNw$n0?v?R|6(=SrcFb6L;jD3tWp{<2oIyFjvD^eNl2JyVMORkCu3R_ znFhGs69>Rb{pw^_Zh;2rXF`MsdPfm%NflmJR!g`O!?k&nmte_`v~b# zUjaEdn}fZQAX19$*&%uQ;oMT_V2&!Gf=XFn!Sbwbc0*`U7*4YWC>Y=FQh@Bcq)6rhR^Wf?DNumNm=XPY`b7 z3LQWb^3tJjZzbY*CFC2S)FX*5St6FO&uMh~s+`)kyliXD22Gdw?>jHazX2tUDchX{ z*Yk<4xf>=SIOhzUc_zzVWbh9?NISXT?5$S*Dzr*r5($ zhJrnnx4jC?^ivkX7-SJV5q9$*>4RD&XypZ59}4>zA-3;2I$F=}|j! zSXzZrh%C6Ih4-y48^X`lr_>&Su|zR|7B;q28lAidYfdiH4xMK^EavYWfNv_#uJXoW)E;$2t%f23sO%! z_OEnEqhSa}d66AAUl@LKiE7rXS&Co%-c_o5hBXU4)DkfyhK$73g-V--O^@-2P_El-K9@2lnNz2TL{TNa>g&1)Z3 zI>-fPEJh@4a4H4!;DPiR6#o|W+y?#FstC9h@y@+`os3fdiu*H0H%x~b@k5E(;QL0a1F^}}_J$eCDoQ@5 z>~TZXz#{Ec4?3|J;o`a4)E^MY9sMV@8u#SYf-`?d_+Zo)vb64b?F9~Z?78ZaF> zf-5gsQMW@oS-2vpu8euv!-2&BYs&8a#GOPL=i_dKByAus__o)HFDQ}A8FxIu!V}b| znOOt7QK$<;D9Beo3E?y_oMjCbM`JlYL`DIuLWMJ$AwbEkk2Okx+7}mp|>m638?u;k`A%~$glv!q)kd9!Y#;kH(FyiuWZD0?22U<^}rCc zkCPnzKzjKVsyt;;e~gTfnWpd`LAV*Z!#VzP@FD!ix%lh&M;nb61OuITjqTOf?@R=n zL*G-N??~2-LP*_xc!wRG3z`&69`b;(au$U(l~_JPtIoQkD?;}8e4V893DOS042Ad1 za+D_`R6B&&NKqp?k=R=l#TJGmX$7f6z34i10$wbGnY5gnBL|X-a~cg%aXwi(_FX(k2A~xnlC&{CD~{{%T`sk}bKlol6oXego3J*+5*= zbFs?Y6DE#2t8|yDwxY0Dv$ zlyQR0zPcjtb`$Koo9tXii?TP^5$*Yyf0ykQ@!^0&qR!6tInjc-qktDPy zpD@29Dpjb@4UJ1n6o@wA^&-QZc8m-C0O=5c&|A%z$3hJc?=BE^7@<1Sp*#@<;A_}m z6RBLeB+OZ?+6rJcCJ}udNf)TtO&U_yS8u{$i4ZsXHD?unO~~|i5G8iPy9Xl_5@W%t zWllazZBII!sRKhx;gjK8s->_DAK%!*V_-EmTBV^hmlfHj5bTk4mX376W7^EXa<%pcK7-= z@(}zsiZl=5nGwY-S<`u@IztkGP5qB?c3yBfAQ(9{=tjv7&>NgTU%1&LNciV!Cy^`C z;;M}<#40anhHPoOigr0HSX$P1hUwIqApUFs?gNCBpsyiW&4$^1*!!NwQyjEYyq4j6 zCASfsO`zc3u-d^5-Uo)bmG262E)vwG$F(ZrYZtA$o*et;3vX-8zS&W z-=1ETLnkxUR$xVf)_qqI077d%E*U~@&kblIqZxx_B^wujG8+5! z(!Cw!9@y!Lv0rfY49XdO2UGrrh^@+A-A$H*{)TZIdtUBd1R_pQkwoQX6s zqZ_uL718gYKBd#+fE8ypFVRvu>OT*UI_01!nc6tL9S*>Yi==69-y+)k4kUtq*yZ0Z zk>C@gLr#j3Zt5C!R=o;#bf_djO}`IOJy4gT+OS(I4QV~C^UE;WxkF9uMa#YXvPHIr zzy{T4T+9>5(;^M2uK;2a-r<9Jmp#MYDS@<>eKy94v?oOpc8YD&t zNS8E7mvpC$25F>2LK^7~l@6&93R1FxG{WcsV|({|KkxeoZ1;0_ch5cdoagyI6|KUA zIWNDgHjXNNWo@`jfK_a8T1ZgM+9=l8uU^j5i#aZ#c0pf5jRYNzGGZkA!4<_FUwq%< z3o3l35l+Z_qH$haz!UmPGdj2OrR8Rod3?;lb79;;KCF(vE45Heb9h9xie#y7b}E*6 zhw0qFVWJ^^f@?ur04w)R9*XH;6u(pn9(j)UN%CybFwQF-#S4@={pf}JvY0ggjc@~5 z0de^Ts|9aUT(Ja8l{Plr7QJU&3f-;$&H68&PtFth)0__rKrrI5`rIEhbLWFQiTTUz z-zzD1Uha%9XrBhg|D9S9OGy<&l|V1+_08~S;w8!%`GSJLg(f6&>YTETMy~j(*uQni zCII;P#*Nu7Wmro3I>S7Vbd-#kQfmB;oX1veat!~hYyG`Z$i&~{@}|_j{s*&J>k{t6 z@{7f-7!3Gfdz)1=4H8R`ibU}AyNdn~l^eo1d$O)C1bq`ea*3$TRGJ<(-<9gy}3zlb<9Qqo;G7twfG;~s10hAB) zoP@N(2E#H3luZ_Ovyh^pj)-1&AUc&D(}qoCX`{HB`^$bbPC1KGbA%R)4q~QVZqOcm zeF;UMGLI5*P43fjV%TEMVrejLseoA!_c*68ztv`_iOFwXTZySj0`@C8eEa)PM1=9@ zFZw1)77XvbP!Ty`;EjLy_e@!1E7)f z2eKljC~*r5rqKj*AJHQp}JZ;d{XpDv2F#Gz8kw$Bq8j+E$FcmDA!u53@CEv zp2J;{FaW!cR+IFS?s6(aJvHaE{@dk-v=v?9sO_bvb%q6!ql{#{ey!9sy4cWsy%<~O ztJkaQ;-rEslx@xtiGhJ|mom+Jg4`j~QDgR3pCN@7i$ix``>uYD|6-aAUUIy?tbR%J zmoi}ZbwnDoteW-l#U^dSzux6>TO*!W*XJu(LXD*fmQ$v0->a^&rq5uQsrkO=LV5>z zij`~_y?xY+N7T#}gM|cGSK>X6NmAftnl&izM59hH+8?t$ccUTiVo6Uvda(oq3Z(>* zb8fxb9+SY&PeDHIjY{?2Ua-Qj?c8=WKLak4D9vj7Og7T6B8v zMvEIwRE6)n@p6yMQQzFZLkXI>t5B@(VP+6*<2_NC^&S-5)u>hIo#VO$f{iF+boUX<`%3Wh167~MdXW5_01c&P7 zS~vE?{MJT*o2nKD4xdGDN}g;pz3Xr$evJRw(q1SllgpT-bLkVrA9e@AiA zz~{A~Qu$$j?S?n?RgpVxAN;5iWht!#n1_bmh3I0DmOCSlXoj$GqyAXg>f)OzDZ(a3 zf2pwL?4mDIrCvr*;ILdW7l~w}e17b^#AD>aDm<@jN6X+{`R!YYla2GFXg;8GMJ!S4hd?mYve17L=n0 z=0_oftYxH(4~gT;%sJ$pIK1E5 zm}OeSKOF@;s)c=y;m|@2KLLc71%AYC4BRZ$+Qew|I`mjU_*Z|H)$xmCd0PUO}f zDu(&2z;w=&vKE8T(^#6-F8v;bF(%C5B$xRDgW$UGgR=22O2UE@AAf!vUZ{;er%3-$ zK=6tGs0>##QAIzZU-6_`<+rgOtwyJ9`G#=mx7hVx@1}0EZH`ZIAE$yx?KJ!ZgTVQG zMhtpj^;4-373olHpd1QTyUW4cmtW?VwksedB+c4h3 zEb`iAVm2dRQ09p3%JYr#n*m9RMW@hY@k%7{I1MB8Ek zlvV`*g`hu>BzC$6(XO%e#Cl1;5yiptjNkhDwAH$obo@A{9o|&L$W+n1AR#O5@h8lA zPs|{Rc81YS2RH#lU5VBDW^s|8x9F8ihdtbZv&8D)uo*G(IADrZIHpNGy2|d2vJv4 z!)x*gnSb{oX~$KzwPDQ8f(dKaZnv9&1~D%?{hF>|E@{)_-0+)^Xn|G4sLi3k%!R2i zYsiZjg=enRVHXfeCmi;j-Lh-tIM{!!`~tc;_8JyO{m$S7P+^?w^)!#Mb4EubNJ6^A zcx^hqd62|G!8Eh}@)Q2|tWEbI65g0UsU(wC(nldqe4?OdSbM}?{e%$cn*m%0nDnS7lCKt8R8rWK-trZO#?~zfOl5WZzYY9T+>) zo%7^t- zm+Si*z6=}~>ljQ!8eK%b zhJa#kWWS{!7GOw)mA&wz5EH5*MD1_^ z@nuv%#$wnR=~cFmmuqE08n9>g9IChRm6}CWjhu0wYHIU%2-WSPT3J@!Ct@cbY-DY< zr&0dp)qYI6ba8I9S34s(b%_f$GY!`&e5tCIM#pOjz2JCM@3{?dpLZbGIch=EPk`Jh zAeo*zijkGqNck232{QKBWg(SmP4FQ_<6e)OLCk*DMr_)OcFj=yH}^MHOF}Jr>HW1Q zgbU#sHlUU0!#`2>0uuEB#xxfm*16PbwE~B*e_-goVMMKSP4?VxC-?;VX=o77FT@BjzT=%{`)1IqtO9xcrXBz;#J}One<0i47NqU-*M%J67D(6ll`tNhjXvwV$78WYTlh`AC`m z3n>pU)Gv633H5E53DPom+lLQ1b_&d+7YBzSu(b|EC_-w}k*3%7Gi!{1-S-0k2q?0H zkMfrgsQ&phOGMDzS+#fXHUBb>o>UGRC5MDxq2N7xXWd2N0e9skHpuLc*>9!>eaHn= zC@76H=Q1{~-6NnJ>M@S~(I>fW$4NR;_eU@~kQw#BrPniF{S3*U-`|HHtuz(CiCM+k z1DcY>TyJBP5imEnou<@;65?VUH9TW?Cy$LH7cZrT1tiM3F0iSloNCNm@mLdmuQHt< zlKPh&H>iu`jNc|aV7Wc}wINRWm1f?CO+!P&oCVl_dbZKyUH%jA4zTP<5z=rnYQ48_ z`EsF`GXa6(R^|N2=5|-=*JMHdfNjqL)sq27g}E2O*Lv1jeDH#t8F9?eUia1JeWi_; z{TH=g0{c~TV}CK$D#Ifnn@oVnuk-n{`>b0c+rE?kK7JIPX^KmP5vVM3s=Bmv2vm1<{ki zGw=KH{gcTC-nb7&W}JXfLEMzS$O=`QrrOk$Zw>>9@CR7}xX5i8WfpP{FXUnJJCwQj z4lwy-{vLS&KY-;>rPzF#ySi9MK0W5w0uDM32BA$-sUP}TYc4q%-bO&MA&JD1PyQzs z@FM4Wr|ilTK*>pwHkr993+zc|i&36l3>&hC3&AT@MW zaa72zc^rEE$(|tLjP6Y0g3uRBctW8Bi z4M6t!Aggf-F7l%_d1OWra^VMH36@<$^fmD`GGOvXPxq&*#hT3n+2s} zCT1iLptzxiY;rOFg&LmS#%#tuvmpGif?O$^oldkfe5rXLvM{^~w>jq|Pxo2SVjlh# z{=~C)MRVmxIs8m{bD=1^!&&A%Ar06n>04r)-uAhiYZ0Trc5APQ;JS+fD=(rkn-aF{ zB1=k6@~HlT{-_1~&o+A9?^8ldF=Ob=mihJ-2=NXUG5d zcejnd#~sALwY@hV0ejhht4S0bVOV)Z|*t1OJ?W! z0*g@fOK{`t?MMrf8Be(?>h4z7TPo0Zjs$j!G01WNw1}n?Df%ZbJ0U)_6fjQiCXkxx z1gq&?5uv+EeuW0$f&$|a>bN!r_>Fnk~?kH{-M@a(5X0x7GI9^jGa(5&#y zlH7d0tk1kB93BoG7J8X4z<3hc4w8FHm3us0md51~aV>1_LCMv_sxQ5FkRAtvoK?sl zB-@!O0=v#c_FP%;7GKMWc^)nf)S4glyE-SL;=|lNUzUErUiC<5Ifm&!j(Cy9#KyR! zU5+&I%qsmnlWA_O0o&{2H!Edl`8m7TViPxvFQ1mT&JB0eTzwRS4kmt%+E*wByKG&5 zYNe`cwCqs|0^9D)6-)i-ZQFH{BEgiDsR+@!tDkB=X31tr=pTn4e!#h{XDPVamauCjTUlT+?uY z4TAv9WSS{FE^Lsv5Tt=z%(ums!N{i+zrJnY(_>;YSugCa<#1NN!KY80Xap~2jhk*+ z;4S@P^4H(s?Lp?}*9i6ZNFkHYz1N2bSFHSiLUG@Z+9lG98H0w;Pv9-vBkMmt&M)@| zVkikTEho_Kj>Wlz()AJ6P*<5c8*{uc@kz8S?yJ#Z`(`*ky@dP~xr9n(Jd!FOjKGir zpo9*Y_HvxdR3CY>jeHa^<_i~Y#YVPcu6SInTba&2v}2UbakaAT_YYnrVp zFoIU`QtEC}f(mz}cU-`zfwYdvj@E;qC9FD94Ze1#0?~t6E*?(2Rj zInQVRxk73^w>bEtDdTrfHF6LU$&=f|+wIg(n%4DsOl`0Mg=jj4r<)6)%M=^Ehjjuv zlXAb@)vQ=MO3zBaFgQrWDrrR{WPx>OwpCDlA9wdNTMt8Wa(yS`Zd>Jo^$6sP3Ts8Z zB@QM4O%4Y!BHHc8`L45Ff?BT<8Gw#%Y-5?o3@Su{K_xXy@n`!w_nzkRJyZ;PZc~;2 zf$mEizEM>g>WC^kP0eHMoz08BKNhdI>`7m7 z+TNcp)Kp~50gX@)K1R+q>3@1HN@2&p{2yzJW4chg_DUHyaR&S1Q^n)^H7uFEj zGeinM0@PWzg5{kD$xTTlm?e4(=da$`a|vsA!TPOK|EwJ7@Fz%!ImcE>%5sBSkZ#{* zVaNQ17*|bY6(t3jI+wIIk-Y!mt)D$N&RU*^`Wyz_tUm3e9 z{FC=jw&$egdd;Qm2Zz8suC1Li!wXcVjVVpVpPYK$_K*wvwKo4AK2k17^}DBs8|f(5 zl!(VI?Lv8XNt!A*KQvs%^T{}CO^-tpg`U&hQIEV}55c=%DKcm0D9#ujvA5P|r@v0}le{Q!32;Xf%Jkn4yh{gr3KscL?_P+S z;K`xhxB;sFQOdqFtaSEVsLEUUMzU8IvDk+BiP%l$0TjDgkNsf${p%5}TiEE0R7|>0 zq2t9Wxms>xDl&68mYQ%W1H#Lr{4)5g1U(X^LLK3&`oJJG_&a(C&S$q} zUk&L(s{pw|+3c(Ci;dNn_>;|YuBRb$i#8fgzWY4ykc%wF?9Z4XccImK(lt>wlm{tQ zkWp6}ALtrz66n7_2f;*WZ(G{@z&seeH~kvo>vlp?VakTjhkW)85BW*yobGqsxMZjT zAqaO-a`D2xPEVDZ&@IsF@%pIJiTJ6IVF|#$3qr`k8tKPbGw5NYAEWDf#zuyksY#^v<#|918z6C{O2E~>qAUt_g38;A3wg!UpyWCB1 zVS2?1Itnl-CXaI7NXe-=kL$`0O8K%(R$3$v5^=|=Un(%R%Dr_H#mz!-^OONNj+OvC zeO20@E#+((yy(|15m!QBYo+ytpq4t1fO}{q>EErpKe;FmiPmT%d6+WXV*}*mR8P-&u0^{$ZZAruH#nsw{OwuCVDWR61Y!eK zs$7xDF!+zK$!w%tgdPwCob-OWy8?j#DCZhn_U(CDX6u)E5t5>5HE4_BS_nYF+16ZI(T)F{wF=cUTU!4pTuFlSrEHqh%|i z+1wSRGzx;`XGd77c(KejSGM5oO^CDGQg561zD>3o*MmTH%GV!Cr3bRrMZb$ZP@2Zf(DVdCvcE{_ko{ z^Ox=LeZfNtFkC^d+zN>Ia(9ze_`j}ph7~dZ_%xN6G4dmh^wVSs27pw=lFbirQ^A}1e-^;mi%7rO!IB8%|K8*zcZh$hHYnllQ)2Nv zss(`E1Fx?J-Ilffeyq{?A*U2_+BO-}yAAAUMOsMl9b0JdPBIhvi$tGAlR4x7O?c)x zLVk)*&`Hs%d7v4IePTWCu)gdiNF`NQzQb2isBpvq5|VsAOCvR-4+)ANIl!dHAZ@um;^)og7Ez6V>Rb=0 zn^AcA1a{(0ikBp z_*9LPQ`^5;P$_lNB?4Oud&91SQY37_P&dJ$4l|3&_tP;}hfCEdVVZVIOYzr(cj5=E zZP#(LK(Fd)?%!fsDhw3+$kk=0=#ZSv-YPlA)1=~RKBBlFDNA5+m7_-UUCY)Y2dO-c zBRT}EBEQlw?QX6!M|)F#b96;;Y5cxY=XiX_d)jcZ+c4L;D6zW^jb1jgGa&}yfl~PgXC6y_ z9rOisbT18uLkh|LCDv$XD!MIi_X*506Q??>+X+|)V%%PHYsfv@pyMvbQ4uaK>SA;+ za{l*hUxlb7O569y?s1ShQizC0sa-hq>L6RESnAlZ`J=P=u?d*?WIG7=s@p2f`b$C|6n863bIsl?7ZRyA?ApR#u$mff3tLi1CRjvlSmNkvHKMwb?@&tG=bNu$E!CUbA_sp`d_dKOKrsL&AODO<}c~DKQqxCQM!6`6yA(y zjw97xx+35)ofz8Le`Jl!5pyrWAJ;a%KVb%DlXI$-XIw3k+n2HBTZh$)_f^%)J#HhX zRy-{sVICc~AmWFMzksFTQEu|rA5tMb^MK{kiUR<9*h4lRy_>h|XkkKBaa!SlfR*{5 z<+8X-cO75KnL&U#9W>+kggEs;8zZ)OH@g!09%9tU06q>X#+OK$i_YO$^(xV`CvXHZx z2EbN5nK}zbxl|;vMhPJ}?;#`D?PzZmj=?g?<0uPjF4Gv0s`1&+ISS2Iq8Xt$BBzI+ z^(YIp#nI+b>4L>DxkwD4Nm(3C-9l44*7C;m-x~cc`0SHciuOf7SBMB{!pj=_-(sbd zN-f;7J>L1~PY-UWQKaW1Pf0;R)xW#*RhxupE(b{7rx(l-tF%P)^EY&j$(%ujP`9kn z?>eJ06jgJaTm6s->q^_!!}6ak8A#3El_gvLpY@g?VvP`W^pOl!;8lA(Q4?bdfSQ^3 zV)3#_;tMrW6`Ev7>nBZv&F=~h=`eoZ>Gl%AC5s7X8@{v=fU}u$?haX^lA)rRykzJ9 z9oE1no$;N2PA%fpj3_w}k;C{aTFZZtH04?yZk!wQ1b*x+-$k0iaDbnHYU-|ECYF#k z1;X{$(q=i&z4?LhqnrORPC^ z9&B5${X`qm`h>jK2aN?I^Z%4pQNfwdA}_S7>lYR5gFf|Elt_>BkG%(~PtfMpceR zEtIL|tp{8ENQbl}*dv?XN3YB4>E%6(OL_om^!GNQ-T~lY_#gdVz@}Tf+k3mFm=L6N zfBVJzCtqCR8`O9tyR&o7u=BM+-|eycUlt};apzgr@m#3Jo@nc9deMNSQaw?(rDr9< z?c|<1xvajuaLwL_>0H^KuM$3N;*Z5QU#W`5L{JgDeoVS12Gmmj*1J%q!mC5B9;y|= zn^u`UP_a#9%&*yx&$ZsjvHE3IdkL9RS3nPX7+g=}>CXPYSjL|95c^tA7AQ zslU0>_Yjm|ilCKqU?jqlYhvyk;Otjger1!DGXI_{1i?+7BVLLW^bbKGx6Y6{#gYaf zis{xo&+q^skzedwLgM0;Px1yv@0uCR-a)G@=W6@1XX!^XP?Pl(SX6V*<)!N0CLTCe zh4m(OWvN3kq*(F{&|U^{Il1`GaW=8JjpHF=g|UO1Z8KD z3=&>{lfhtAw!Avd@(pGe4tODm55)+6W|k;7<4>vG)2RSzF;)^?Qd1Ec1-GMeB|#S8 zhe>gkcRErkT-#{APDgnq%5KC{JB7rZI8hOnnAkfR0pm(LM0Kx%jdR@EauqW@87n1z zi0J{E+IjZ5xt;a{4+TQpCU5n?_JlAtEPhGaBsV;>I8UUa%|km%I_0{Rt(S9K1jRB)%DqfFc zN;YVje0Yr)hn$uB-=r*`jV|2ebW@)M{0CDg752?AE^BM2RoGf}4MqQNA_Xk}gK5Eg z+ap6(q5kGUM3~^Hzp@qX2mBuql-d}fsMIZ8zmuk(X6tftccJnFjKa=ugP>| z*GI@bZzrFxQa$>_>F+kqr5_7@;)9+8LGAB1f(qkRR#Abl{9W#=D|{$kN6+bc7iON= z{d4c{5$_OFk|EH?X6pfJ?%T(x%Ny)a9+n-^fQ_bg2wPJ1KYJDI$r7Y)pw&~f}*mt`@ZyC?6%7%vT3c3N@l z{q^}-qP@hkBU4x}HZ=5i=oW?Aq4&~FLmv6Xby<_RootnIKfz>ZX1}CI!Y!gEdN^J&kPrFHD?BfI1I!)MUHggTZbLb)0TCLl6wUrFb&X+?uP++fpB$F!FBHJcd(3dD~{bPGE%@Px`>hw z^uBJYFoD5(lgDG}5@20*5$kL;aw*2*p|#E={N}7r?5#B7lTYOIsI+t`q!lJBCxCQ+ zKD13h*l^N3s6=g7l%SP(%Umu^l>63k4_1CfdQ`Z~T1`5Pbo^QVZCe_d;R+{>9%KZ5 zrdtBIpMda50g(%pp#B=)X7k(byYeJPAep2~mVi|Oc=r_LOqg@|5_AXO(bNxSD**U|9=+pK7JM{3d099@igU#2%pXc~$v zy-fg+dZUParB7}T<|iu8W=yP}B*H5|E2$>Y{Q5Ew2i23QMx&VP zmi9DuU+d%5O?4<*y%30@jrqmDw*@)ZoE!|m@=_JI ze0*E?S{o#d2-v0uXSh@*C07rch^EQvi~N(0V1DXIK~kyXu5NjfiVHVOu_&lx#7)=S zpC6Ip_(vR^l_GO6A`#drq#Owhkxw#hD`mEE8HLg)zl8cQREN5)lH~s zQrz2*!jqUv3*$0%hkMn~f{T5r$0z&Ymn=>@7s8r^c(iz|$Ryw!dPUV#mLU?Sq}bht zf`Mo~Na#zqJ&Z?>$LcIR_(EDPucABY4x zDt?0M*_I(G2M?b;nCXPfkOAx+wZn-&oOfzPeLVVoy118a=gOan-4>(ezdZOvPv)yw zx3N$OBLE4>l?OSGk3D1{plLJ_FJif%!mSZb=~gt$(p$uiD2gFc2iD@*T~z_T{nnQo zm3Buxq>4Zsc}VM+VhELr zhbYCA#F@`{2Tf0v&zhuQRA?huR{Vx#rvePb2JYKLj_>)=Z@y(7TVv?t5=km7g8X7H zAV&_I_Zyvky$YIpK|>!00-rUegy=4nfvspWY~k)d>v)mx&tx8p*{U|kGmr)&!DQxJ z3o{GTqrtImJd`BOl@yn`A;X4LG%_(4ch{sp_IYLo{?qGt?H7i>CKYpm!>>+Ar7_)` znyJ&}Pm4TEs1;tkpuE_IAiP7&W!)ZJACFf7QswEAkd6xaQVEJRek*d+|Iqar4*rTNTp}qbJ zhv{%Vvde}cme5M|j5tOC9u>NUS9sgvO75vEJ~mX}ffa%AEn$>Hno<1xd_2mqX2EYK zs^M`3xrLC$Vbc&vR0Qm9%nu$4ChR3n*zye(xg*8K5Q+pNWk4`vWzqAlNvMlMgTX0u zD)R;&zjI7OgzB+&Sl{?87??z3(_4HeUKjo>G{Es`F#&H_-0Uk4DTa#=+8ER4bZk_o zTX1JnJjc<^?-5PfjB5bjf-UbP>~#qoU#8}}9L5mwdQ9`;|E7Q}`%Rd`$dLTVp9H&q z)~-{j>F~;W-}x+LHAa?sh?j!-i6~3G0`T&hBM?KFOHXYq&X!n|gRqe+ zzirGrjSu$EXE&bz9{;LmpU!{cI04{3fa{rt#*BUnR%wJj6@%b8w#-YqOPSK)X{q*@2-->6jUemKw|C(++lMU4amInx zfj^66NuQM8cUO_We=U`#PJ?6VyiqB8<~FF;jJiD)qX(kT&3Yl$Od%ed!1 zqRu8RRw`*tIKcFWGrHY^aJ#C=#lq#9^@XBW+(3i{ej9%2jMkk)6_5y7VdD%pcR~zeDhKUSHaM@(UDpzy zSyJFum`ST!nNAh4I=}YGv6ZpfkCd1LjyW2#c#@)wV?KGRrQeMTafJ}Xu;!jcz_ z@fFoueFcoIu1ibEz{7d~nq*Q4tURg~NVL0#t3o{qNI&0_{{lZhcCSR=z$VwZy}r8! zy+#KCt3DJ1y#DkJlmUZ&NzWoq4r}4m0cbKz^l(*F%3U=vnIsYiUDxOJOM-W0W(-g7ekHH) z;|M@97-R_A?KnMxq3WFf=4H<_Wg%&Q>;HWFNln4%3#B$gD{F#+hP>u(i4}DiQ^hoP zjsSjRJbh2;ve4k#>$6{sF1ka>*cD#P{MReB0Y`-6Sh{bOt_ZZZ_;gC;H2e!apxhn^ z`c{IM%E0}=;j zYb?l=R>`~r>ab>t2nv2i!oZD`f$cLTou5zj&c-r{LkMD=jSfCyZvkw+s9?DI^a zy_Tji1VsZE6O3&Lj}!lRKP%K*gEIW=Yo~jtCG71tPC?m|BYg#IUcg@nKVXe z6Ef!LT+?^+0ZH)a$WH=%sp#ZjCmPP!@)`(5)e#2GzR0$`%;wgo&}q12baqZkTy#ZY znD33uS~`EkEzKuB1wHY0)Vq8nO(Op)!UOE( z&ObeKeuxV9h`A&eN{m1%#_#qcE6XOS5f>AvPwt*?1L2PuK;*FIVUWbB#)D%s;*R@} zY+x8S$M44(NjrPoUT#e9-)n@;=U2X=Xe;dUje&UMyTXTS>~~sz_J--BfIl!+iHJ9U zf7iW?j){=9rOLbH56ERcdKA8dm5&k{l;nwF=&BvzXkebd)|Zk>Ql77MVjS@_shmiQ zjjcagNlZ@kLUCJKy}f`P*KsdwKoG^<)oqwd=xt#w%I4KC&nOX=dVSihxrpj|fR347seWVdHn>t)@QP%*$q+v&BSwh%X? z76SpsQ$PV_c(a$GL&(>QjvH({eG9|kFSj`zhRd?cha4pMKw;6h0qpxQo0AIF<%T{Z zaumlCQ)A=n+ATNwiq!b^X`yo~7336>e+6H+nM`Os>r=O|Uei3Fn*59}5YL5f)2-Ee z!}v3Q{?lO$2^=F*dz@ZWC=>8fFD%qs@>06BrXlSeb)v;7aAg+$C;R?;lZ>3LGvH}5Vx9W&pRdK6vS zbO!y-i#<83LH{|!5n-HwrB5w_F=Dg8%Hh78sUp~Ul6;a+Lp+_>R7j^)tlgBD@};ce z*_#tC($(^B4!C8Cm{#%;vijsjKF|Dp;POwLowx9Vdp+7wJU!U`!7A$@0_U6*rT+K* zdRK;pMC@)4<_tVNwg8cn&gXZ)?-z0IF+e2h%@3G)SN3Sftc;yRK6W`-GG&8UGT@4k z8P;~!`7rn@n5oyp^L*gBy&Q;q|L@@@j{2`!>wsFc@?vtIh@D~xgxc+ zV#PK|-Y5X>K{hla`mAzaIjj1>t=u<#aGDN(8*SO|>)2PMI#_lpu8c$jxBx0waZQ3x z-e%su8;mp@jC@}?iz$97DJaDt<;Ju_n+OPX)+)!d^Ipn5uT|_LkFskg9^2h79=gM; zQp0$7e1}9V8CrWZEnHLA^V`Z$xAiMCrvuw|o;z;Be3gJ*FNDU*el8Bu!3+I5uK->M z-WSW{--Yb=6a!{FWsJ`pgOf~QvThtIZw02nX2=X^;dx`)gzVIL2#v8IU%d}z zTgFP?3O{+aLJ1f?mDcGYx$SUqK;k~jvpN54@&M< zl@BCf`zuLs3=j_Gqw_{vy#!on4&)AX{|K}F(GywAGw%9lXv z&SP(tL#ULZu)$>dvpfzdx~uA zzcH6Wbpymxz(dVwNfD$V8k+@}2&Qm=w%FMNU(4?$nSZ z%O0rG?XrtRBm1-Lf|)JynVS&kfXlo)nRi`PVKJUl)~s^cz3?| zS7H5<6^EfHgF1%v7>^pK*I}j;&BIBY_G@a9%|rSbi9(!^jvl9f9lH4FL1JG#T+tv> z82&?1hjxjy>Omj-v!i#O64AJm1YO=)4?49okkp_Qs}hG#n?=O?Xbh!2h2HTbdKgzG z_da5QmlWpoN{U9K4!JdMdV3Ti@$9(QtcLOXm58fZhx7Q${{j*1ZbVfSRDW{^wpC{A zn*}I6TrN+#qnu*YN3&r!e|{_dRPnWVJw2$RUiGDl#YOg4CZ{npC2>u9Uisb@XXAAO z-cXPcclKsNdOC_)l(-A3D845y%!9Hda)jHuK?YU?#u~wb=!qw%8>`?uIHMR@t&z7l zk8JNwOJo-g=AQTJuUf?~OW40}2~K4uK*xhz41;pu$5CHu_FS}iy=l}-B(duY?0Out zL^H0cI!==Aj9zJ} zg|2fu3doIoSdSaXMSV?Is2pqPUwEm~sPG$f4c67wErE*pPEuq@aVm|n>lf5t&;THt zT|;hSNZ1~Pzw&bb#F=Aw4Gk=IL<0l#-g_6j^pr1=IFvDs!A?f)FImobB1QN$C*NFY zgs}OhH#IU=O6{b@gc6A4M7}97P7Zo%}bmZ^^S9%O3+gq z{^n)`|)vTKQu=E6wbe;B<8At582c}30>yeeEeajp1mciM92S9oZ+NljH4|y9;O8=dA_i-AC zO?!y5VEvu}eBOmQ`{DEQ;CJG~IW@l7Uztd%Z`jn+^Hd2NbA|;}Kk3D(#hLqG(Y)V_ z`vqfTSFlgcErJG;f~h#n{$sEm+}%lfVc_ow4kWEtU$A5V2D8a%7vl{ z;o^x`fQjlzc)#pym)azVg#{N`g$kFDh?x?4baL@HJq`Ma&HnVHp!AQ{Ys2!>Ju-Uc zV_ZsqHK`8N^KizKRFJP)jjfAdeLMi7q=M zYYZ1N|In;)4T_AL%;*cD2_H=dULIX#MVNf66L4VqT^9)B zq?FGA8qFdlfjgYa!B5bo7hwJZNV$UdTVU}|2!R_rey|59SPBqs0}o&tSSl^QbLoe| zzN|Hsiwx&FoyMc+2*4@u4Drp){zY8qYCgT>CL^XK%d((#!-r#&=PjIL=B?u~)x(5& z0$O^~LiT)bJ7yK~XHDD7a4jiGFf8&q5X*e%x=VYGKvO)LOn@@pSx$iY`I2Oo&x&(rEZb~Jwf z?@Pe(cdg0~=Z3xo6Tvek2VmzO$1C^skue##0ZFmQE_&kuSXa?VD|UdDOz$;6#MW6= zN7{RJHsRgqLJ%pcUt`B%pD<(k6VHD7XvxVKI@|C6*Rh#{@fu-kukIBeUkWATh2LX} z5e^9f(DiUkDx=>LbRHM7sw+WkXzMYXO<&{;TW9ma*z7%>g8JA>#`+G^d(lPkL8>P~ zP{G1PtNjb$obgA@OTopl6m5%h!KNHlD=d=#_dj&xT?5*&_kA=&B)A{@1ySrZ=k z|MQ{MVb$gUlHi+Dt&-DPO+3C1o19~+s>)hTYMzxztL$dS8M_0@F-y<&AKVTXv-+`C z)p%KQsw)VeSEQDtGWo zKA!Pzbb-PvD}-&i`^$gd0LOLL@gu%^-xwWQz@C!5l?`sKvLyTkifvz}fB&CKS!^@; z6%^`zu=Aq$9t?q5P1f_K2Ev{nFDSOA!D9QgN}6qt|7(%)GuE~H>ih_RACDjmrC~R& z1rEYGB@k<)(aa&f-&5lk_izy_bgc1~8oP(1*okFBmBVY})rDzpB^Yhh6*1Zxu1+v*i_^G$7CW4#^ z!OFBZVKj>URNv?oU^o9ePrwh<18duyuI(9d`>#C;gkL=aLs=*=sijvAwgLp6+aU2y zxanLGCtlXCMO@zVlk0wW{$H8`+^_Xca6MmbGM9MsM=DlKPda^H ziiDv2lwr@luqPxFJd>rr%(Z2Cj+vkBqQ!L{krA}FQ2Of}d9l4qlG(f=Ij{p6gE=3V zF!2}wy$G&ENsR;esP#oIw>xWNAl!~`=vrWm$679hrU{zWx(hT$J0w!}=-jQn2?oJI zNnME=npf8y6^hh?l1->t?w3cubJ>xBObN7K!p2)6wNnYqZ^rw>u@Au)UA)wyUj#0L z;We9`pCEO+*C(a9;|uol2SD&a8s_CiyE9zo-(3u)@WGyZ@begH-;MY0j=wv>i8CjHn>{@qkts`^0u^_k!@jprTL;B2@zRAbw$rrj1D_Kbg)#@_=4@?~CdSAozC zgdLa5R;P6?`dEl@j?dG()fPwUzqFFTnLWx)so^%N#rgvu#3Z)f=?y-PxR*)?*SuH! zvR9;Xjrk{^ld-{P-p1#8jQu}T1L78++iU1MK2^B2S{jhcA;X;|n$?m_HS!()HrMDw z;5)#@MU3-baLnOYX-+?WF|QDKiMRO}wT$Tjz^m}A%}kZBOW zl) zes`xoALjw}`m)W){3fRNNk$g-&;uWfIhmJJY$8$?4h$@DcViM&9q!q4nH^bZsWdj#eaxyR#Hd*I}| z(&s|yBJZm`L?r#NE=Yg0&0DRfyXmg!$aU_IPA~4?**J|+vi}amfQFXx{gDg^MU|Z-(|d; zz*{87yN^{M%||O(-654aWdj+CCQa)j;7Y00{*WnR&qHFv&c-f}v8w#^a8diRKP-uf zv?6)lrKid15|zwtWcYM@!pwSkohZrJx|iD{8<=smXjIbmK|^_utDjJSG-5Wd&qxQ=|RtOh4y)lA`M59E~eI>TBYM|v6P=lg00?dg}OKYmuK z$?F@)kN*4v26YwjIyZgyOH{uzg*+T{E3&6nLH@jlYaxnYLUv z1nRz8?-;pv0Rf{uxe}Q3T1UVUIaJE>RuNG(L=<<)x%7=iOC3{Ptwz)U74+e5Bz>xD zZVY~OmgS;tb)@aqci~;lT+zkUSE$V~ya7ACa_?HcWT~wyMEMs4R#Mr$s8nI#O~|GDlxHO{(m| z#3rHapeueVp-^W(>!_jG?Zri3GH?kY*nrMHI0lkMVkFr`kLeE^tuQVR0o-godkx&5 z{$(zz!U6w9?(N00)-vb^eDa;};nvi{TNw1Q_@KO4U&sG%d8z8Tl#l$*H(E?62l2>0 z{d^g3nY9S>&0X5N$Vhs7arr?jbkN`)WanZRon~hQ#1MW2jz@wLWY8iy3a!JLs)#w$ zQtsH180Wr~_5fvV&shdoWlZI|7a2d@=ll1ehivi>L7s0-sOU7WA4ljZFg|DX-6?|> z8uq@9QvA#t=g}}hu#UZ)eHS((8YaQoQXrUNJMx08w}FI7qOfzT@2K&<7Kb0|Hx zpzQiW9?Ek=U^Re>lRB>n*T~GG(|li5&L-HE9tN#>I0$!lx?N`%c^4~y4_vh=}+`4IM0CeWl%@Xfi z^LsIIgy6?V@@#$}{w)QHdabhl)v?&3ufM^iPyPgE=OvEa-*gTS)VGn#yS+2o9%W54 zWxAV^;m(8G7ptk{Ncj}oRA=fUA+ZJOXnok3SnnK&th9iQ8ycx5FezPmo@2c0ajI@| zT-Q`BpbPg5_;6B))CuJE$i_>Xr^o@@W8zGl&MWfewrcO zv|_&r2tQxB=fw}MZ=}!Im-3QxwjuwrXM@} zojDOhId4sC;d!>A6O}v8A5C=d-XHa4r>A4|*TZI|m14e~L(C9G%w7A@Pi#hC((q3I zMjw>JF%Q;vjpih5c(Bq}rV`KSAR`)9&t0wib_3~I}%SAz4` zO4Lng4-PokMeJN$kuel(gO;ohZ|5s3b*s%fD?Pr5ry4aou%kwkq-qJ^UmuLeepTf* zTg{N(0j!eGvbt%!=#KcKRpO1-OJ2f$r$U`+4R zE;PKsd0GZ?Y|A^qy&drUbIx15``CG{==(yYk=CE79Nyf3D^|#n>c45z$;PH?d)@s> zszkKnRf);{zv)>TSs$1j@+Kh|hRC+E!FqSMpRX^J=~LXNB;J2stYLlSuz+%pZhcsh z;zo(m>7GxMiMzqyb}Zk13B~T^R_edQha848C1R}xXq#4t73(5$SZZvuSB4y}In<;y z*PMS4_-&SiPMhL^`2vk7)aPZ7z(Qx=m#?VR_(Q2jn7?#m{b--X#>? z**ED+yp8#9t3JGV#Qesigz^0$Id;K3maqJQ4}xe9F<)rT`tVx!8cR5dr+Oegn1 z|J7=X9Q8jv<{zNdn&Nq@378)Rodfw|`Pq@}iiiDa#NGO8Wq7m_D9~yC_U`AeE$RRr zYOTqh8+h{`t4KgFZVDzsujAm)=-yXcle`=OJ#fU+16q)L#{JE;J2yJP?eOgAjrN}YZbDq7yZAkTT{U3O-7Ks*WCh0Ync&KK5I8It2 z#drbOfK4LssJR~NmW(P)RsIB*48ooB<^hE;`}6&Il5RS&Pu&z0IQ|J-!p?e1rL|U#p1b9;f2U1wFiZpVTB5}#!ppbjbzqHj0Xe>oy6&m^ zBCfHuDPXo$jBwx9RK0${4?ZbVT>#9AI(OAx2MZVayue`(zWMFO_J?t~AY`TTyBg!Aec;Ubf-b8D{hc+E z&abTY^W%$q`fGUt?E^)B_#^X9t(6PSf9pY{9N35>3>aAh`|Itib#UAwv=b0g9)=&8 zhb=zXRJ}TRdjC{ELYH)8i5MUh*E0_G!0crkOcUC={E_=(8=YiwUuQl{2t3 zPa!ni!GS=%poY~c-;>pN%368C{okL=YDOk7&bT78$qL!fzt;W{%}@(MCOSD#I^BJg z7F%0`zgu{y9SdV-?O2zw?&?P;=DqV`o0zw)?@H8GF~Ao&zg$j*u=ts#%X2lq*1}N9 z$5shUe!y?JDYp5 z4%PY`P5&$n`4@RPB$z_M-*er7H95E2dc6~jeEmqCJJBn)ojeCgOhNhs7)0bT+7wZb z@W36V6T zjno#ZOtbqvq~a2^-5qZ!U>)#t;O|(+-%-4V6)m?PWwT4uT`+ot?qBAj1|N^OfOQLp zGp=1)Bb#&igMWB!xt4aJt@)!-&Al35skdUKCov<@jM!i{3FN7=3XIS#zmfhK9zZSr zv8n5JFIx9c(_Dk&so~?%Hk?sx!v)!t5%%MVDE4`1@Tdd!X9XPL=iU_HE{>crErC#%zh%Pc9WPUul#! zSftpFDc`?Gz!Qiivv*eT+^R|e#zn|5Esk2K(R;tvCr7#y( z9B~J-x$X#tXFs3z=?Q{Sm=clOQGc6bQs+fCwUu{f$Mzo7lEE{wUs^{Rqp$~Z=59ci z2pSFsU2}nds>n>F{z0srk!7lFn$p8j424xUXjAszi1|7swVORKwoHU2ja#OYvnTr6e z`%x5S<&BHk%kBK(A^9X?6lS92{2jnL)5ln~oz7qdLamZk-H+Z5py3)&s51(~b@lWP z>EY1G>urKdyWg3w4$Ikd6IBRodhVYbv_F{Pp#G)%38n_G^5lZ=%WnrH3PkoM^GKI06zV ztr^J|L>cx-R0Zs4{8r5F|mhzh0t7CQS?Qdr)h~I{ZYtmWy-R zH7`d^$EN~q&Hv}GeUPXXAv<;qfpcFD-|$P-2K?J-PvR#bv;S@^nBC$8#L@ip6qi1fn- z*z-#pEC4Hz$$tv|*ES~~sw`$8?RTe--QQ_p&_<>@;ZW0Il9-rypf{6=wIEy{;i|!) z-+ySm8-ki>T;BjaHOC*PuI(AqH&)OO)4l$}G08W7cX#GKopPE_Gn&O8*Kt>n#n&j< z5Q{S83~1wBRv!|wBOA1O$HTC4|7z>P%B|HrYen(*Ws5HY9=1wPw}m)(KQBn>ErV)$ z^zWy0TeSmCubG#^p+a;XHr{}^D-$x7CwCD<7F{ffhftkttTI*MmpZp`ML)3>8xV}e zb8f#V@Re_}zgyAkvtOMS@LoOLnbr4st`6>Sa_w1bcSPT?d?~wZ{kOu_`EWK&mtbC* z^vdgEv(jk{w)wLCSvpD5*{S?PbqCk)G8da?eVi*R#JN}GcKyD6e&Iw}n2>q3w_puR zOEfQKaU-RqnHpkrfMA-;krqad}S5KqheGG5-ZHe_p!;_y6{M9(Y@(BWN z<2#f%#W|;qD@dkC+IK2o!-t|`8O8f&Cxi>plEm(QU>coK?S|~PCFU zQ@dC=oG`G^SbiN1Hbt^rGOhO0d3l0Fb(4neoVfWmS3ve2q-dXkp+&`dB3WJbuFP+=?5kzyuqm*9+ff@05*<6(tX-h&)pCEy|OwDW@#$|b2z8v0%%U3{u zgoqM+#KBvZ-u5TcAvvWbDRdLvEJ{)PR=Mpv&hU zlFn&D1Xw}@7gCUrOk!p)rwvAegS0c&U6Yf9nz&vK4WwjT(koWUImr`0{X z=n$8=T^|gc+U}>%qLc&h(%oneF0z_0n^fHwMt!=a3{}{g%eBSjvK?eEaHOalnb*ug z<^}BR>;s9PBF+ZMABvDA?S_Y^*e(yO(>@d+J@sj$nF6-?x*`Z#|Bi#cy{si?Po<@O z$5;hUhiD)9nGYj`$K8a|J~j}CQi5W?&6p^Evi&ag{Q(H>LuN>8W3P>ZL5M42zo1Bp zHWannv(o>AQA z8(*cJiqg;rtg3t}(Jl%yJyMce{JO6kdKEETR2<)qmJdO4l(Q6M5>j4aQ6gV|s8U{o zQmi^HbrPuEv!JeYVb4uVt5LN}_LjQUYbWwz65$rHmYAGA44;avi5{D2s{puZ(kDa` z-EeI3RGZ(KeZI$L+hF%JHb(56j&lMOBs-@J3_ zy2aaK>`8$4gugPR)c1Xf-DkMPQ?@PKQXI+Mi16`60%hz^#sLj=)019p%O~HtjTI;&l{0S}4~x*{WIcJQXz5 zAH|8(O*JGrJ3=KlWmwE-$h*vnii=a;J4;9?Y{E4`okRi6as>t779f3FGP*hH<3bFu zX|>nsYd%TEYwf?dzL9_%=LCo{nxzcWD43)qu-hr^BQ1kWX6pDQy)N{*OL(;vug@4v z`n59zJpbyO@xO+z%V)fWv0n=J9yR&zwo3!GQ{>0SrfZ}l^YX+u8JaljF&g=nVkpQS%?#Pfkaur!7r-5ojf7S&8xv(Mx&q*mei{f^T$tC`VMU-!yclugNeZ$8tKIG3ci zCgI8S9h;z_N_Ak=PZRm&xgOYE?e$|??f}(Mfc;iidrd1)TzBsU&b@4J7Sk+gM1n}o z8tZ@Dd`aQvb4C-P8CLB$PP!jDXjl}m+aO%s3_UY`;?6fcgfwKH3(d+B;4;bSrhklX zZxPCX^G$eQYSp-XiO}%MP{szyZoK*N3Pu;mg9?KH#9Xnkv9YnI){GiuLMGGCdq2`P zr5XA!sNo6T7b=>Xi9PU_3uP&Io8~-{Bq5pi7!)}+%8IvWv9@DEhw2k8{L|T(!Au$h ziXt8p;^qL!x}&m}SDXk@Rv5dHxaj`3D=&s7@#WNCJ>hZq$7EOaBfb_IacEMs$PmWC zs`>5P2~Rv;lvxHMmT=CNbcMOiP&{u)J&Sn#Fg(zoz_>}#`zuC!{=Oq^={A6CQ8@#0OhR{c;`hu)yYxYq-<^`K z<&RZ%CCy^%ygWGdnm?1iB<3I~U`0LbRUS*%G39GJ;$MF+8^|^9=oA*-kyb)=ffn3V z*7hv&U!RPBGdG|Q!!D9{xsVg%B>8fq28PY3^hS+8dAPdabz77oTwPD_53I5T+&S!^ zJX>?0sPu_UymDHgsL9AIQ(-sb7Q94QPb^lg)b`R_W5dXNu5#u;=jSPDOk&DR-?wK4vTLS?2%5$=F)BnVl)cv=MS^ztl7x9+=XK?W>;kcRL?(U$ z@$fMtk_aMj$8bJ1{U2?)0whNa6RqHI#7+5Gr0??m2V4e+3v;XnJ_9mIbnz?|J1)4F zlBh4md)5TqQ>D!q;c)()iKh`D9ej=Y;mu$!U}{h0f3I1xzhlR>STeAP!%Isl5KXpj+B6UsBVe1S|B*{kY>m3$Dh<7JSFC z!h=YzKq4^zdr=dbK*pQ5QOK0|Vjfj{g)QH!X7p9R%oS>XD1?r5hzoL_ai=@(<|9YB zW8VE>ONat>SgO`}a^;&}*vK~1QCyP4s0|n~wdfh;_B>r0HU?u_!VNi6zjD(jf)G_l zLSNR<(>;Fc(eBFzbDPayW#^f>Tq#Yk-e=hT>raDyaVJN4I?DEv`SV~i6nh?XEFW$D zs@l90?^Y%+sv>2l*sSyGC_p#Sf2(!KJ{`3aeCc6(G_0f&v2}(VXEze>f^f@iULL|# zV8L|&AgDH z3Leb<(_VYIo_I4<%}0WHxVz_qL1Nl?^gFnDMhwy4#D)OgpiQHm1xu-M_TX)CXqAj) zlUv@7r2C6B8&@s(#N_W+;EiBhyohnJO~pDaCBBW0WO>>|k#lEzaz zEa~e3?ltCw`&T?FDSgs!MYx-ZuDshwsEWHG;R_8PajU9dUj)MJf0^9Hzj5zfXUlV( zbj9K^>ZP)yKE@+JguA}*BA^;zPP&_X)V6G)1(rqz3nsQF*Ts1g>I;bcdFB!cO}i+M zBsJHX!TpUaOAb@*KUcV$ajxfQmG^Y1pV7_wU8h{8nLpGd{4#F5{xgO>kx7+E9i?#I zBc&WUOj;xg2#u_Xp9LM_syjd1?B`1$4h4pZ&Q|dDzZI1~>|Qc{2Uhh)stjp)p4~)#0DIePyPe53 z8U8@~Nt`Z_U{5p{7Y1X;f+vXnAfp58GMj1)l&L+x{Od6l^X-jw5>fBZ?DgDLd_)5}BJtA-WUJ#iU89dKVr z#9x-U?>_%YVE>a`o2vXS7^{3Q6RSjTghd@v79DoVZ$w0HTGK+swoYWpH{l?f$TSq< z{k0o@a2}lZ948w0FO=Z-(Ea)s?#UmVw(eG)j>G`}lqiBGGvxjohFDd*`wkJN`TL)! zsE9oXQ%(@yctIJ|9rD%hDk@Kl&rJR&9aQoKQIL>dh}{RJLqnOb92l3>w8i8}wk{{` zn3S+jmsi^v{693iz3O(^48M`TR?YPTK8 z+%gd+b>@Be*I^F79!9_1&#rbYE< zMa1Xd;BL%RF7g2*s07D*!^|7;89r1e=(RMxof2ZeKt$|igBT{C$>%+n9f|Q(8n}EN zM_x2%$HLw!vy?*kCAmDtdjfO2i6dc^0`5aGB=qY7jhQ>tDp^LDr8j2g~w0yoA zFP8_0K&eHndCv4;2D;R8n7NA{<`RV=VN+l}cg8S;@kxA2g+Ur9=67KBNhU#;IMEk9 zBkiOF-3HZ$O1vn4e&=8x8VI(@d`34i>X9W($TM%A*8=g%H(sEf+6hf&hp7{f~pJLqcy7kFHxR$I+`J$V%u4m<=`Ta@# z{vAe4;GnaeIC(eM(C9_^FnNa@%Dx*Pjx)9%USk3CM|b{Y8ppMkC3jIS3rEP_#I!+> zTxB-$Gk*V5F_pKN*_QXdXY+<#kk17F8h=Nb3<<9}LhkryOE<16X1hoCt0CMwQ%bs} zsoF=*(L;zHLM}!Qv(|Z#{EohH=IrP2efY$1t^vcer@}II72&_Nrg`zPn*cj$55g(x z)eg@r9YZ6}7GN}V7Ky$=hTCCJ3lvjEK*i7sGOr-!0R~=T8t8H9H z*h>Ar#Y;k!)Cc63B5szPR|>3%)l6kw88QJC9eEPZXN?R5s*86G+I@OW&P}~LkDOAG z-;cyizsbz0{%}wO_>z{3y5y>v%@RjGWN!kiYE6@lIuEi)_aUbrwu9% zhTf?cFBq7*F+HJL&H~(3{gr@|;}b0B3h1K zdKu}qFa>!u8hN#q6V`OFN9FtCxm=;fX%cJ}d1L(M2Nj~yw#xW;#7 z*7`I4Gw-IG^*+v3+vGe0nQKpHwLRT@07R30ch17lU$LRR2`9k>v4! zn)Wo)Ig%S3)wdiv_X-IkceznS=iqSdNJm#G&ua%~pd)%2z#&&=+W-T>VsKa^DOnVM z7o=eBG5DDt!ET;>{vTlb=>nLA;&UT$aJLXx6BRA~L;ME0^E3Fq8yQHueez5Fn8v0Q z+i+p&OcGmCe)h4zLch005D7-Cp+}Zv*PT8p&!a<(Ncx_LGu4iV#DR*(?re^PME~mv z8A_^+!u|8hREGOSSF1B8R@)BF*{}KO$qe_Y4X-H@8p*Pn!0YYd`&{MfsqDvyeJtld z$(r>OQvrVn#u+dzg~E0SD?-t6zy1m$--~#t+JyJMT>l>H>=s)6s9%2gmOOcuB`*6b z6FND?B}Ba?*pu|DA->+J7_7GQT;j202MH0F^Zq^WxgWKyylK9>&D}370FA?O^WoJ# zP#@*>OG4&4ZMV`_@bZ|k&{x}Fat>P0iKS$}MAImkpbyO7D zN2P$>^@6uj_Rtn4aND=c=u#i$&-ng5OaDkDGe7k7fa+z%nxae1RMCDHKo2aDI214Z`uh)%Ztaq8Q6=0&sSgsoZnhc}>Oq&x!07;t$LH zgB$z3YpWk|u&N`*@!u9OC65RsD70l2G&Ja*bL{XTxI;N0FDX4~`*meJ2Zz5zBt$}V z7cTK>GeZ563#IPRGP^4^Vn2x1V2Jc>U$ zibaUOC(ML<5lFQSjkiu94%s|X$gBeuj(D#d^3|A3okGw2sp2*bx_cM^4A@@CM@xHA# z%uBI|&yo9dx8k$GhWLh#OCiTc{=rRZUKf8PTGA5er}xoceX(x5tv=)OnRRbVM@sSs zMq71x+se*KBPQ+Twl8w}{{ZbqzSzEx3Nq(1VplKgJ-o7w&JXq^g3lQiIdqbGnXEsJ zeqcV7T{Z>KB0%I#!aRu{El&PsP`*Cd+m6P{vy0_GaG-``?%FxT6w2J{{~L; z8t!g#(=c6x`AGnx-Ufs=AAU%k;CpVSAZXHz^V6EI0_NnQeRjP;AX&o~jVp~Oi)wkb z<0O9dC+6}*8Sh-gXUzdzYxz(n=?JbOpCzZMo%iXU<4aC+szG2RE0CK2SueTqiv2(d z@w5kx^M{&llFmhM`*!gX;$JjiLq^F)f<9NlY&Ud6=7o_f+Yb_`N5M?$Pe+7dnL$6b zF>WBmO||wPmYAjf&Bwy*nwY1K6`f=wmhv{~J|UT}WiNyGXv4#8#g}l10{`kl>`nhB z(JCQ)vR(n3tMEd2hMojFj&sv;yM2>DMW9Nx{3>dhn}zXUjC}iQh7(Zen%Nm4qXw~U zsu9=j52A7QL;hXCJf;*Dll)n<=yeEn4J-Oz6XvE!0%&5Sw_G0IQ_@!i_!;PQoF05W z*s=dZ(n9}H|HthYr`YtGfxuCTJT>AE(29N5ntJ=!0M^V+~oR5o)vSla0eHS_Ul-^&Cx< zF=qvx=-mVTtNFd_QXD#~L|LmJd5pS2jY-a#qBEIecDaAHRia2$e60y}t){Hh2H;FD z&Wc0k$00DilZ@9MY2(_3s5L=5PB>KZ@{o<#D`Jj;_K=@ftgNhVm># z%82fZ{YN0M#&EB-gASRZhofNTH1%`eSD`w{)O3k$YZ+$SHN}~O6s&rhtBd3Ue7N)P1Au8sCDi9l<2APJkyS}H3N(ABZVI=kuybb71s7k%DG)!OkK%}W4t}=+wD5?e%qhn!=QippLRmuN&CN2R|~tP zwx?5m15O_7x_PV3t5hvF8s!ZWHo^#GR;bdqggRTF#ZpO`4uBIs?m!{)7;k(=Xj3#? zvdxdN`n3L9z6q79CBf0<{oRw4ECoSiavL_URQ@ANNBCAZ&So@}rYH~(G8evcWpD~7 z&j$}d$p^q;x#J+#3;s87J?LzBhPvosR_qSw9MVBO(x*9G<2qic&bANGH&~r5iKnfk zxS~3>+lbq|F1E(yP$AAcL5RB3j{(}s>oK-Ajs$V?v@&jOG~^RlLhpK?b|1ZMDfNw# zuR)e`vjp;yMF-^{wsp3DU~2oip*wSa4M+yYC?7|T6!J?Ns7zF!x#A6rUO9yIUHND~ z=R^F#B$^exSLiXrUGV9ge90+--Leqe}aT@LMuG1Lx^S zA+xY!^fG1-(gXBY)V><~fQ#ISTUWH^lMwerK`<%bceCV z%_+rUsVIXw%x_pXY^K%z#GLAEJ)x-z0mho!#M00HlJU$n2^NXEj(bDs-gVJ{BD7aI z1NVn9Cvy3r>vMGu39T9B1LJ zo`#rzG?oGCnz&o?N3BmdMM#5hl0oHN2Zo#l5t`;Q41|;H7)f4(N+{NZ>ypizzn>j{6AT}bSja#R^^?dG$dmVoB1bL>2R2ay_~PSS&T%kroG>Bn zxYPgC!fNSRY3f`|KOuxFS2uNi>&i<`ZbQPf)E@*7_W0kvd5VUjQ;lX=6+l%42*%aF zPezYom%PJb3DePmIR!qxe^rV_j^0Wf%HtLF23qP*cjA0AZ{DrFNpibPGxE=X-<<8C zUZ!EUF9oLA5=i}Sv<*H)2U@jcSNGl)@hC-f= z{eh1ZB8s;&zawi6cBhc)e->bgo?b@f9vm@>8TsiE*LE)LTkz`dk`g?UeoG~W5c+%2 z>b@l3aSmCI;xfVu4)Z({k-Iz8?!x2XCzpME;aB#D!(JR-V1+uRTDdX>z1A<{Pbv!w z+i6J=#_R;sW@X61(|rPO4^Do2oYz8a}docJ{9w{vWCE|JeH?S|UDz=;1{lg=`^&-Nol?x*`46MG** zA307|S5%_@G7Ru(GWcST#4(}NuQvH!5^u_j3$-JoBRv(jk;RIgfGvNWN6wn*GnCVcsmW%q2j^7jTjm0a@ zufZgVAkV&|8bUR-t3*W~Exsv=cFy5kU$v+_pF<{>rKOviA#CGcu*yPvA5l*tXX1KO&g8+ zS-r*-?h?G^H-JcplEN@^(?V{3RAiTw{3fllH;DqDXnlaQ-W>Jz?8r-UXzgN~i!tNK z*xxIRTHeF|4^{6S*Hja|`-&jF_Z~V(Zwd$!igcu@APUk2MCnqbgx(3gC?#}6dQp1s zpj7Fd5IP9~ga9GAdC&QsbMO5(*`Mq^voo{SteNNgq!|L&?V%7VE7q+oud0aJ@R8@S zpo+|9^9mw@&Uz(5hg_!7irBp* zDf~PF5MFi*>r+Y_ZBdn%PS-v1kfb#-H@IOoQs%PMq42D~sVJ#^_eRX0SYoF;6o33d z#mlnBci|fk#(Y`Tn0=wveXN3~Hy7}JVGDn(=TamW-Xzwmem2D5XqsL*p0;+zofx9n z(-@M%taf%|7(U^5+*D9sre#fR16p{5U93sqmW+@cUi(fE>dWw`2oHwiTlpnR^+B%jSN9Oyf!o=dY;f5}!ptrxS9Zsv{_0gGlOKKD*e-;mx*IoH>2MCc}FMalpe|U#xwY9H7 z4&$GE%(sQKZ8+EZV1$;?fH)I76Lwg*tGX$yd|9m9s<{ulvGOn7p_hLsIpDk``(X`r zwKUsyA9a0ac*K`|&%jl!x=jteXYXK8U1rLnllpsP^L!^g@$^M5j7K1mX=4|FO`V~Y znm9m9sH~}s$d4IOuA4ca`!RyL52qg9N_0E}B58Dq?wn~7ScuYO5gE17yKxBk6p^51kb!hSZ@Oe3K2tL?^C7JdvQz5?qHT@Pki`mlg?me;aoXU;Ax z-_+Z-qGL9N|K_?CGf;lM=NX(b-uD4 z#pc>W#c%N;V8=MM%%3mZ+xddG*qvTQ#utqoE>51~GXIPQYGEmq-}dUo0=}FA;5SEZ zJr8y|Ugn8MN+e3QT$-Lj*f6-`!D}Zi5jR@EE+eq>RJ|UTn-9tc`%|P(bo==m)ud#4 z&Uv>07w!Wr5b}kTDIY@yeC>8oFB$`^FgeQjAZvQYjGn6mdjH2k71!7p&-0&OO8r%M z>JPu1Dr|1cacdhrX_++;VpCg^M9mm{qRA(*|0fe8%mk&qT5*>c@$J5{yb`lGWNm9h zVKdAG+5-?>Z|V7~Dzr|9NX=T35YwGLIlI?Hr4B|Zu!)}ICwJ*5S5B5|N^6X_s zV|R1h>|o1&az`iT@TUCo_Puw0?-FFXF>|#gu%kolxI||}`>Xs&wZO!76jOK4;QUG< zSzN|WXFjHJQdVCH^B3xRkosXsVZ^RJP>d}S^o|rWkY?Dnl+ox>rA-2lY3-+A^)9wB z;M|6cq0d6^)@ngPEZ{Att>|>#ZWz)qJHu_#2gGq{<~#p&Z}a|5KY<;j91DrY5O-5! zXX}DaelE1>xra)cDe5UJE%aE+cFNna-3s5t6I8l^8;Ypmo=7B&3=bQ>|3!|9H1bg?-TRR zMH1|FO$o=yb0%=NPj-D39!UD)tu>lN%{tC;yg#en2ErQ%9$*AuKM`Yn66lizJ`6#X zBTG(S6JcK*JhJXlF;4J3{);yTCpETb!y3Pri!91j)NIiUsk~Ld%DSxob&y_UxUUg3 zm%CJ>er_15l zGYSn5r`&1>zq88m_Vx%Y|<1_{_V9=+GNrO>%Qnb(kbev1|SDU`VQ2w-0>){bU|6*N_yy^LmjD#imMvTuqq`jD<~*mL`qG$^#GC?HdXfg2`VgT9nLRCsS@jn zLX}t!;gSeDn^E@v$tbk^sbN(7Vpx6%6^~vlo7X#69_<5RLx#pv-an+vZ_l_MNTbJd zK@`fR-TIU>ubriFdgJR%-P#>)j+cP=90W zRx@I8{psCr8CRV0i%DIZ8#Q0_2M?rv1EGSD=Kgm&B~G!Tr`UL!Z?fKQ$Xdit#CKw8 zuO+EIEXWeI@l%Ixp~h123dbUEO*2$@P}z$)cI9~T60S1!kM;@A_1%VZR-Q-7m8Vv!GzF4a{XT|5M<#l~G^!(_1mf<@HcE2i(%%#fM*< zY{)Yu@ZQJ$FZ&VzEJ>)YV0C%0d{Y+OB>)~<%w^if4-iFiJ=(|&gZvt6Ltk$>UX8ZN zZRAFl6;qyWEKh#eAwN{`+N1WLu~VKaC$3w}vmS6-`jL@yT?XD7x&Wp*@8Q^Td&+nE zzADUt72(zdF{;AkkR|Vtpd-=*;7T*DFxF1R2r?@07adj$>_fn)%sx!i{CIZE6pi)Y zM}**C`pT^P-RXs6v$|IgmZ4}k?wq`ics3ALlD3*+Wfm^vkzrf<7y}62x4F7YGxW5= z;`|B0D)0aN@%$Mb?fert*|Bm`*SP}1;;-T~(*8E_(WB{~`>*EsS)94nK-jyrc!`ev zOZ4)>YHsVR(V841%+?wCRXoJ<_}j7C1La&<#S`b{c#3H3=r_Yvbt+(P&q_AN@ydCR zQ7mv+EZQWx4oDR!mYuoEjEcP3Xfd|CD#qSQnZnAVsOOjJkUpM>iK?$|Z5+Of(|7wt zfObpYH4V1TeB^?n`iIqagqY6;!Y8$Tf0(J&G2j17C#pQmd%@D@^ZD-Ni*;Bw>~kFn zd~+L(I;^afegtH(fdlX%d6bEiJ{0f0v=tQCca}Dyrr6H}(etVOoA+mt-BGuo!Ro`{bun z;=8Y$XcIk7#59;!y6_XsRsG?lg_R*5qB)~}*>Awf{w2)ANg@X#mAg*&g&}bdY^$P% zGCT|Mb`4U!HW!Gl-I$lAkdB5;S~o9qZc?uX#}A=)=hN1aER9fG}Gi)OwABXHJA%d9L|n%Zj1qK^-Z zU67a@dHYaD#Z4J!fh8MORK`utz0!_z3Tgu>p+Z)0d$Vh#LfOm5_2pI8UHgJSkJ`y> zZr$rchjGi>%M+w}>A;%L$|n^q{HkQ(-m!lixIcgQ@%;Aww-<(rf9{T~uzmbXiK`Xz zNj?KHDew{mBw|hvS|`@|#&qaRYXLV>Hhq@-nHHSdHEW(%&cpfBu^=o$?)6<;F&p?~ zJ2$ZFb`84f#QgZ1wZ=(~+aFJ0kHaDk$dDSaUl?uQJba&7lTplqfPXz7p6fNq@VoPR zH^=qj`}FIj4w1w5RSySX#@QzyxA2>;_*d`4pQ+~kv0ugMT;dhFHO0?%>*Xl85WXRv z?nFYY=H`PDq*r4lIt^;u2{H!0ip8B*+xjgTB4#6{F&(mB{cf*7m=l8`;U>hkU+As% zcbwf_jpD^rzr|#J%K&wj;QKh4BKI2v2rB|S#rUlO)-G4@2xE5Kd~sG{&4^$l@ChmV zof-J|2mA-07wEbX!5|pIvYF1%JL~RdVdTmz_oUrId@k?l?WZfuEn>W0WPlslZ9UC% zmFE`xg8T`nZwyLE!P>?P-j6^YhQy~{&*4WaBefxwwVi~O84E3WzROE_Uxq4NkN4|G zHhS_t&wpKffHZ9W0@H1NJ3bF@fwqMRa1I=v^S^fRCG{DX?OVjmB89SX=!uo5VU}V8 zC*nrOdBMA1UFnBlFW8elPbln-IE*JHd|9Lzq)h3`x|o|=3Bi&lxD^kBJyz6J7n^im z`1ZvLw%C9=$5STqgW#aog{y>>>%hJbw@fyEsNW}zh%TeQ5<|MbO5t_tEA6;_e~vv> zLr*NuDvQ93$@XgW{PkBLo9rw6MGM&Pc=f$Wt4&H}l*;Cjkncw4Rclh+eGK}1()ZKK zFNT*19E$Uu3P;bIZRbd@r*+z#R(#y#IGQj-xh$Qt6YXv72+oI&PikM+-~Zlkt;+TL zi&q(244~*3yCnJR?$!$HbI$bJHxDe|XZ+9={-6nq%K^&+a4fjBuG{0I^#H_EoG$Q6 z;a$j0C!Ik^p~-&JDX+-(!g_q*?`A)YhZh6}K-Q55?OSKMHwFPovDmIW&ahx3MZ#)yhH;s^%OiA~3|2+;@ z1=H~~qNT}b&Ol;Pw1PN(`0kQp?iSAU=eIx{I*3fOsvBpa=!ux&E#~ytxSsHh=oYkX z{_9wfaUSfR&mOLdFCt+8yiQN6qm7d{yB-hjibb+fI1K)n6Ir{mK^JT}t+rrs@oTl0 z!{>ECq57HrRsPepe+!m^BMQ6Oc6Gc*EX}MAtg5O8H-bs2mxB7ao6Bljp1UlLZ7Kbl zErWVMccNoZh~j1l{_1l%3Yzc%xuhp!92-A#*zpEsgep#fsO0dT+IHfPm+e;F!Qe69 zHwd%pcNF=oN5UIl#)I&sL)w4)jVcZ&oKHH+N4X7*&=f%UT@9IqlWvBL!I~_bb92xp zPFKaG#wcclX7Y+?bYY5zK;rbD-0QY>=%wPC1jxVQc4Y%oMksrVNwEUbD#=&|Yu~Xe z1wQrZ`{^{TChl~yP+RLlt_(UHEnoFK?kPq)1ONh2+$lPWuHE4|r@x;qG&ty)nV9gn z4=dk%&-pjRmdxwok)Q=Qj%#@f{_+Ohdkb$7-xyqdL28EzQp1;h)~NghlbrMK=ExNn zyqNC}{0}Q!k++D^CA3IpZHNUbCD#Fq8!hd5zDpW6t8ESo{LMkvW9Oz~8XU)=!X%Ouy6OD7^8~!Y0h^~w0&cM)rWXI?BBO4*0e2%FSree) z#PD8cCg0v9Ysuaml;EGazg4O^$`Ka0f=yGkYG7i(VFjpdG);ilRj^7Iev)ZzfV#cA z3WuAv1Jvdpt=G`()|y$c?AOYqTdlQyPHSx>)g9uc1X}onalG}c5>Y5?%z5OJ$wt{VON0O=T8iz-Kl#oMosU@Fg`81QAPi?GD`uE!f?f7Iz# z4e5p+S{zooxl8y3Uq``*wuC8bZ`;bJNgGcs1WINu%>81&B_Pb8SS}y?FsCK0jg&Ny zQ)&riyQTNyC5sEk z)!miEAFkwow&ysh{kE<5CIB48*jGU3C&ZW4;3!QiD9IQ8E3VzENri%(?Rjd0-B3a% z5<2nK_V~C^grNK3*r@^Iw4*egwHJ{R2Zg>UJq%J#v&jd*=6u6`I(E2+ia}XP+#!j( zr;CS1kYNK%Ni6JPyy#<;0xw$aViGAjfPKq^`PUpk;~N=sY;AZw%=BdhVKs2B_PNy} z%lQoAkK2i?<7<>f+AGc-;Uz>apNj9=_$PH*TTCRejjuFuSyhJF=&>!Bg54a9SL;>8Aj3>-@=yV2@d~( zl%kzoDQ;O=cJEVeBT)ly%l3S(Pmhj|n>T`M=z^c+1`jzq^?MuhU5XvW^F~KK$#>Cz z@aDFk(o0MAfOQ*0LtoU5TH=Ri7bmGM$>98la5Jeyk&0ETE&l{w7nxRGf_MALFShhV z!IsGQb&pktwuN;oy5e4SEGjx+YmiApXRXD9_bRJ{wVWrSxDg_jm@har=yg9sH3*eQ}gT=WD>@c(CU$@`W zESN3CJpu=-7C5>HY1Jdcu%1}(Y%hgW)|C03_9JhKDZ=egsHc* zah@$y)*T8>ya+WNN?FMFbycZXZpFn=x%MFq#H+NpE@1m00xv@SZxm>*}}0Kh5C0w`q?LU5F% z=)*~cEY&YWxwyESX5eEdC9b5rkAimZxkli{x$m3y`FZwJEVR!WC|ah26N-_xc;vE( zJ|R!0^wDVcz`7izPd}JTzqhQDhzv!q2j&QUCxSXveP9Od$J8n!T4wVWPo60>gvB|r zHpF0|sv`Z-8xJ)8u_-`1?tE5SGbGQZ{AR8f?bdK|QC`wBUzJ7~#|Z+lS8Z;fqqlH8 zcF0iMe*W!7BXjsrA6C5=3b3>Acb3{>NCiSdQF`s{M5PzfQU-s*X`wma(=>mq6_C3< zcpzuyJcH8wR)ClG51Hg>SKG^MgqaeOr-(|o>++IKsFC5vbpHQ?7tbS)o z@x@OkL1LA{gNHnFMKwj4hAVnuo%yxOYq;!R zbtsHg=#N$2V-C*)p=JO%&fWt9B!%|Kfo}-O`<`pyot?M=PVxs|i&ee@vpWf*K(zGh zcGh*=9L6-*UZVIEwerb9Nu#Hii*`KS54$m!3*97%6A`u;JmyaM*;8L5imQcj@3aY^ zzTK*;*c(!}pqm3n=^DjN?Os>r=oHC@T$}+tT!WczWLRR0x$fg*e5jWT^MY#@^sOJc zwm=#U+PEIb6eqq7qh}uX|J8Y%z{^Ga3~VgN&i}8EY_c^{QmyVJy;}>k$1GLWNz9_B z4mAW_dM!i@>yJ{&j~MUVJ_WyO!5Qfm%6}kSw|v-e_Ho1z9#|;q2owS^U&)`=3xaE1 zVuL#=q2s#D`k8@szc~Va*L!ySB@qlXJjD{u$O{I=v?gfIXW2ACLi2(-6gj~DE^f@j z2oHyBD^~TInh=_=Lw=lDKkpHm%sfm)df9Z)6QN`;yYbBXQxsSTcPXPkC$F*WV7bHH zGV^$g@9mW`S=I5(z|G3LNY(#J*f<@FU)1gA`3{~Up=6*4A8}8^0wAv>kb_deZ-6Gc z{fgUB9wYnRJJk3rMT;|qm+n9Q)Ry78q7lK3V$FrTR7om!2}>v0{HQO`Wdp#@$W3ji zHzR}uArVRL7zXA9>a3ZKT{LEi8uh<|HgyD7uTwaI8aA`A8@D^8C=jnn+(uCdY)?SX z1ZSz$yhkn=V3$vz(!|X!St$X6BZ5k(B2>LFHno^1$TdE2NQuqc9sD41MSQ?Ah2lr6rZ%B)BgJY<*c4)Y_7lnJpw?Eh~Xb*!I zS+>`l0voNjhbc3sr7Tn!ByU5azRNUNBh#){FxydpX)EA4W`TMAMZPg+f@7NYB&>VPuazatw2o z7fe+_o4CC>%1;IAsLrYl?5T82U=)u}mWQA4yMakb?cYFVhC|Sf%$I5`M$0A*Saht# zeHi;!iULCn+=is9m~2}8S9gewF#6V$mKs(&5X7)!xa_cjEWncRY%U4!0-ER=!!A_> zVxK%I&h)z+GnBS5T!0E1JN*o%wc{cJ3-ZmU(M=M6wqjQnE~D8Kz!f=|#B7QC`xUb)p*D$1i(6at4$ACd}o#zXcecuP+4*6<4 z)Q!Bw<_9q<-hW~}hxv%MKH6^KwZ$PAJw+GZy^aLkOMh8iPgW#mYJ1WM@VO^yE;_~s z=WAuEyq4(wVZo}ie6m8RI}on1WFNFa;m86A&scY~iNq+zug}>Ph^7$Zhi!yx`zp{o zE7ip4j%Xk{!@WXs;#6$`%+Hg_3~sx9J>D*`z4iQqQXfS2Js=y<_V_E8MB6{6UzHD| zq!iBWiiTEys8cfo$)7+??S4LP{^#{mesZoGk$)?xBSfH9Xp+bD?u?SA=3_Ime5Qq) z{;Im(%K`B#V3ZU&MA6!2ISjclGk;Zq3YX4|Gp&Iuqk8Z;DDHh)VQjy0G&#Pn5CZXF1g@Sls zB8qp_^^JQj@|RA)&H7h4ri=Vubt^%WYYt|Hbm@@Oi)T{OXZH^uxd7?<|HfsEM|QQ^ z?b>f94iDbh^z^(ri)LOVQT$G+u&=^=;}hNRg~7BOAv!kCQOGA4lpCM-#&BCaZqNGT z2eS2V6mN=5GBtI^?tN5DId{^*K6kI0!VqzCokc71T@Il<7RDwvR~8EW-oe-a3IB%$ zKo(BtyJwX0C+}&AOdsTKJ_nwxPUQu>8=c%)g>w`-P54q+-z$zNm9iDft`KpGirZ2w zVBedz4#3kY$)oX2@_zo!oxiJ}vDuCM!8r{I&s2<1ddqSWA8lMmNEp^U z4Lp-a__tmz1IId9OBVX;(!hbv;HF_qr`dM_&wYH3!S&_GmR+6?oE+qqI@Jd0Qr62p zyO?=?bi9g>2jSrl6E%cygRlXb^oK6Wy@OjlWY+=ftJk(hyDa0fA@DX4(-guQjd8|? zw38~Y3C4aAXJmw{WC;bEfOKYuAYG%rkB|_b!69$n`IdzUXE|Z+?1(n!I%p$kt;?w8 z7|-0h2SXHxb|Y?c!T(Y~TXQ?Q772K-g6ziknKlX?Q1(m^>IUC)Ll7>xo>SSl=eFwY zOiLx%5>0|@XbxR|zSTP&Q%-&uJz~Jc%J!PbSLRoQQtD7);*J8Vs}9J~I*XM7EK@fg zuPFo9E$U4S)W0e8C~-J4O?BJ|;%={p7BP^|TZQ9gJ0BnQZdN|+v4iwU8rl*1a(X&a z`8GGH1_}z{;9L4LW4bk)rKmPe0R!WY8pMIPW2ghyEOT1{TwX%IPAJjC6gKh%d!%W z2N2Z8G$W&%GjF#y097ViUby8!Xaj#2Xr2O6Vf7x%i~dB0{nA9ANN@Bk?U#JdQ33XY z5WK>^9D2tf%BzNNc*cnp!ajT#C-VguX^%{zOm5YOgdym?`tNFz?euglAnTnp)#zb{ zg%@Q8nIC^B8FmAgj6!rb=@?-UBruW8I~`fLZOL=ww{r#gbnNY?C3ITptwjuct@|>FgN>P!Hr!+j5d#58>^!xiXxTo zLr|dK9r*sLTlY3gOhYi3gfR2J4oRFQlp-)oPw0eGPrp`<70qQ&e7tdCnm?1CF^XyIXm~DueyK$G5N6nKVq}*S^Ky&wXpHS15n? zgg%+9fNz&Z9byZUO>*0w*6FzTnuYP?zob~OLU0cM2w>G5jc{CK1jri-qva z9JdrD3-do7P&jq|)dX)z=C)R2N2ng>PmM_b$7v~Tj(@%#C|msFHtm#P)ZnraH}4?$ zl#O5oti2{z43g*~(9G?(*PeGUql7_@Fi+4{Zq5YXf zlSgJ4dKRk%%DblZmqaeHF!s4vgM2?Qsyk_QzQUj)h z+hsRBeqrA1^NGSV!tQtB{w*=7hcr>5zD&JZfY!yGAM=i%{`9y}!nw9c%~=^1*A^c+ z+gIr6VRmRfZZV@knV1Edsq+{*aBLx|7HGNl+HL*4K3LQNx+iiCO5pwZ{d@2OLM8Zw zKhuyjcvv`ni$9$T4OL}f`;~T}G#1;@HZHCR{a(Q|Am8ib7;;2u^aVg3&Zdw}I{ESO z5$Av_H|N)(S1>Kk&n36eM?J8r=4UnmIMQY-0CO6>2=mMllJHkzOPc07nbkh)qyuMp zSmKAjZ4X-B`nZquOnKt_<(L2_eVwG3qWl1pgu0<|-4|1@uiW>OYB(hp+RlTX4nw79ag8jN#a2^?dTF1lnxAxi;W9c6h*D+u=`!6Hb2!CoA>!5<2n+ zktFRGq0M8Fdo=|b1+O>@hw691WM^ziXImd`=<4*G0L=X(m9k@jY>I7B>q_)cylCfgBIH4bR3?6?Nu6;HZ@Bs9u>muY{;@4Kl{j z+bG~W3B`0u4TR01`& z6yZkLcSx1x0>1G8V4N_T@kX)q4B6M7;+yr}tFc!?;Lu_z=nkEsXQf#+;}Z)?7+Vr6 zFQ_Xbk5je;bapV@K)1V(U>A!je<&}W__=lv$oi>*eaD~?S3-KYbkLUiJ}c)HsuGTJ$$X z7HH*QDf`4bU20khwYoTjye$)drCs{l^Z72WotCo^beB+Z>h-Oe*#S6?{47`( zqJ8Q6S6cTl_joHe0ISoV@9z*L!hWy3Pg;vL?G>|%U5VHLyxGgTx@%Pa-%I9%WG?Q) zuID#*?fk=#jl4Nj@lnhzV8(UsITl@0JZ>-3kn`d`H&*8vydN8utPM{tKFc5T!WgSg z>b1wdWbMQ0mx7^n#=raXiqUtW*q@eos4t*zX&nCC=r@sH_c@e#|3t5SEw+Rx2XUQe zwc0pkt!9M298KLhUs6YzlAOJMtRO&UeojGb!g|^|p5EH{HR0K`p}D@61@blv;BkZi zlHS9V-rcnJ%Ks7!zlnLV&R^zRt)0=3ldQb-eEYUrf4KlB#Y=ngUd2!e{ZL*rLtOU# z>22i?QfRo}yH$gh!t+NF#<39D5f9gS`gT+65;HToyF&gMc8IanQ7^V_Q?EG4LBm7= zpC&krRFzsS7|n}`R2on*pRXlvs8r%ER7DxI5p|YmoR+@Y8LwXJR9$tPvLiM>BVU5 z1D0d;cGJgt)LSOuCnqmpcHe+TexnMYsh2jxgY3DtV*zp~8VL`SeEy3^Oj#&yO6Nc` zWL`-nXkt}&;7{*$*DLuSC8Wax9Xw-ghV{4VrU<(Y`$y?=FMlDkIckQ=8gfWGDrU<@ zg}-nanhlRe`0_aUu;BPB~Qk)4fKf#-l^Q z-J!28R_A)XE@EZca0UYgrYG|)lLdj6JR<289hG54-zu*ifBRX28y@z}tUbM4Ri{tI zDFb{?E@?__)mM4zn7}U!r^8`pYFon$(>M)rD#Pf@d$xJ)RYJQz0yzIQA%btmsHoWR ze`RQOpZ0yIQ7a3Lu^VAH2W&Gp3>xvx^r5844J~Z$ki$y0RTZ0&B7;ky`JWP}&7`MJ zF@?n@uK#{yG4cmJW6BYkqZLKSPz-A>nHx+Ji$j6@f_kWz4FO@=~1D|9(Eb z%6t`}mnGf{4#%$#=j9dI+MA9EnH!WH&>z-y6L$Cir8ErjbB)O3uz=kOm`=~WR_mu5 zPE)k@3x1mD+>vJA+vi_2P#PMoJ7<~mm?lyjA?JVD>R3qdQEUt-e2Ri;D!DV<%FsbP zDyg#uYba`%-N3KrpalhGw%-UVo#kDB{O|+|#jV_HE@?k#dHrq7mQ!Z1IH2WQr$Z%U z%Z42u+{|1^^XI|$oFN5P>ya)12t~44#aO(+xyY}<&olff>2UyY-fW9xQWHhT9}V>S zYX^hko+>%VR{T1}=pcS94Ynl1(T2RFSS?49QyJ}^k-kEuXynwf{kpN%;@>m=k4KrG z>T(V!Hao@0g8*YvUOP$Ql=9l}9dnA{Ed!Aml#Kj*R8)(4!6^(5CW3=hmhVX5gj-KJ zWo!uSo?K^#22}UG4vfnas@5R+y=J8JC9%DIp3AN5lPP>-|8E7apLOAsgFyJL zYgy+ZXRnmLY=jA>19L;ZWeLwMP)$WS&HkXcQD%zEsp{cdzt+O1MQXCt1rAC}ozgqT zYnUsu#YHa^tny=GqYZ)P&`hyf#s_COvy2GGQ9yh1x}_;%R=b}=PxhjzZ{q8klurlN zX6RyJXFcz#&}p~+T}Sxv4&VTO;Fxc!&bT9S3q1MhT(Fg}LZ=A&E(8Iu30Bv?u=f9_ z%bxVuVm}Zz21~A33?=xs2ism{=(AKm#*#LrcqaMkRH*_M?qOH|sA-TU1_$699?K+_ zqmgRi%(H$M0GI%U=IUpaCnZ~(I5s_PYx@`eA$I7qRz*i~d)cw+dphc;!9>RrglyZ{ zTGq)}oq&&8ddDA%Ji}Xl-^a>YB(f-K?$9tCbUe!;>>G^Vz_$V%Q}%#4S*;!(wUXHi zz>{d+Tr+Ex_0~PV20EXKBtOrPc1^FGEaoUFO`2w0|D=j9?jW-<9-8zMG8aN=V-fD* ztN_m4_M5rafUEeEJ%2I%b5hxSSjwuE)th*gnn~6Il(5EBNh1g1MA53bKN(8%cYelq z2WqV-Rtu>Ji|nj3GM{zW#bj^%Q0FwNVJ+v+iKVf;F5niy>F|wj0TmKTKKN zIZb&%;_yLbGU89(Ls{Ym<+CL%KrQ2{yDRjmfJU<0PJDB3ZFILR9BjS|E2QY zhj~?eVc#fY&R?O^dt%2i+w#eBkSLmP-=k#Ah>I+B3gm65z=+~U;L5<`rojny&e?y< z2?ubM5#E`mVD;ZL;o~9_hJ;o>X~Xx!BF`TLGjwzJAxNe+%5E^)-8Q@T7`Har2sa2X zCcnP;8i+mhw5e%hBvLYdy%(%_O+6@|=0*3D9l~~fry;-Hcf%9NBTY3(gH;nbDgds} zKJwflzFt3+7=FRPHQRvz3IFlGm|-f!R=g-5lQNfW>pI=`%HGi2%#T~Ik;AApl%)oa zZ02zJejBd`De$tt7gfPhDYgkPA4U^R=L?rp-A#hv`lV#_V4 z*z>f1tuUjabl^s<`0@OH>~n_hpjToC|`ak$oviAqq6n_{aKIZS9_(bvNtI%#K10u^Ms<1Zjd9Xn5lp|Y^$i|ECJu{7gX){&`jya zcG`K<=J?kbNA^Vi*jZcFgj3%Cy*y4JCU}+(6b4xhl29lu=iU&uW{LIE&=9JUZgGrl z&_`=G6FdvfM_AFqc2OANK4XcGZ=M`Q(A|fU$^S9&9a?^PE#;iuMYA|F!SF)ubHXY3 z&Tz@r zfFoxBQ$$tDw^De*&lhKQGw$0mNu9jtCsGAMKiG`_-;a(WT!9RvjwL4XV@Ej6#c>!1 z!ew+aC+d$D@0IlG_)q%wF;5?26swc`ar7F=TH!}Tq~6T=!N#WNFpJrXcy2mEXYslQ z#@iqEDJ1^o4W7T)5rRcCtYLCY%+#iaJx?G^4FOu=Jx<5)&9K;fPmfj-^YKEr$}w9a zav@*XVnynx{i+Z)$NGN?s8UF-eT z!QAlLqvM*FO?_DspInAZ6lfv-$9XU8Tvh(>eGLLnWzRf;*Y^hZ?>9hRl~Al3ro~=` z5%DF#(m#@rdf&{;2k7FFII%AS^8E#`9l86d=WPW?6h#?CD~ z@Akbg?t;CT0Ys7{3J~T{qT;es0OBA8HgzWWb+F`VD-AZYv%5f8rEc zE(6upt*0NRxL(O$4iEXk6nDL9Z0}qD!xI&{RP=VLP5CH_O7iel%8dW4%fuH&MOGSQ zD=Oh3r%p_9za#otu07S+8#=iVtY0^2qgS-%q3725ccI5RHv+AX9Icpt5cbK>mdm+I zqBKh!AK@{~_<7pNC+TQm` zlNd~~+Y3m3q+OZbmE|pt^l*-E(jt!EJKutwR%5RpHQHWEER=2$L$dLec`%XlbZ?|+ z@$P1`l?VuF;Mx2|F~Xd;W+uiwek<<}s5qYcekeKIV;v*N-R1uHK;d^DKsL6bY@oaBWX{B!bM!H@@%2poN?@nRg!5B!GQ6v0L1P$th*fL`;-zq(ypo5Z`nET z>n|tt9Oh;`aW%+GuN#4ugws&@;wTIDH{U7AZ#_SkX#TI+lZZ3+OmJ*N9?K^~UPgBN zE9P9He5&uf{7+{f(}Yd^C8b`0el7@#zGcV(wOdEu%Q#Gz=#%w}l_RMv51=}3VEz~5 zSj8%|oHcIgD(hG0uA}4lPzM^v!hK^qjK^?PA=9PqznU_~KD>CQ%-yymWVZUNh3gK< z*=Ep^c;Za(e*IP+(kEl)wjk&9hpeeRE+z$U~607OhE*uD@u{%lL+3B_9TYYZw z9UI1=v+2gi`uNG#_hlWOr9eYsRp0Lk9>tnZ6V|Adx`g%TQad)1r_)t07SV?4#)9mh zj3V1B7VA>&6vvZpOZBp#*_ETT3s3(ERuly1;J& zLKR3k5P+GBp(*66cdEtYTMcTz9O(DIn|sO;`wSLTLj_;9cwjlA>mPI!V5_1}a6l{a zpPM`rpq&&R?!-3ENUzsbonCsNlxv7|Q&SU}u>IPhr>~jR#OCsf<3+8+ePDALy@_i_ z?@!-%hnj&{dymP7Fa$aKbBoo-;c-<4U9-sqJH5yTn3J-^PExiL4Ik*7+3XO)+??rF z5l507O|Hi$>0wp_ z5JB)hGLUFIIBIl)daT!?`OzO>Iy6ye{6yWnEmiw&fOq<5`5|fQXR5(m^J8ol0a_?d7aD8|MvX8GNYZDA<`vDTv)m?; z7mL28o_bT&fz0dV#rlhCVdBwQlDQ4fX(NVFS}h_r?YKMX1d}YO+g1P1HB}n8YnY^| z(K^(eIEn|TM}^0SHmQKWH!C)uYyMQLwb6dbWZ{>5nz~zqF4H$5;o_dSZQVNqN=kxp zxqdjl@NLWeFTb;=aj-#6;cmUc)2;>($T-l&IpBkX31>T58)OFGxA(bc=LF&;Y545= z4Arf0X3#UO2V4B98a{|Z((&=yAhQMF;s>Hb%;Eic#IMBD{%Cjlu%{UtFJe!xs1Q>l z_90|Fdxy()*pCWQ`oq3j7TV%qJb^NN?rB?9Q&`Tk6EVpi zVEe;r3rPc6#4xv47FAy2!xil>g`t`j@37Iu`i%a5M8DH3UeK1*S9f|uuplCBXr6?( z6AX#tAWPoPYUn%jmnGkm1^?^GV!$9nerh&(1Vf51T-~MqED97bjG#(_;f|JnRn~C- zgi4c5@?j0XHp43>K9~N6)@n0c`xOp4;r{X~Q>YL*-?E&>tN5Hpz$5lW6_VuQ$43ow zo_KRsNkbY4anDrmlUT6|rhLMSdSTRpXJB>`_>}uQZ(tHtL4+S26S8)Ie*`z>S=*>bNai+kXOmcpOB zOq*Z1@{hB!*KxRUeq|*dw7a(#O~s9W zY~+ohnSzHgM)R5thIqp7>RRMU1^w^)?tfO8LmDX!urBtG8200yjzHRIB#gw_tDoxN z5z#zj2;eerbDjvq0%zT_M7j}H$YXb*I{im|-P(z$yZXT{ps=t&%Zxbx6PwTg?Z$rV^b$Px{IU#jXC$p!mUDjSG@!FlP!d<77cd%S;~~a{Bl_?c@j_{!^XLqiD!Z z+DoOlrr;k}Jm~-0yc0VD!nHKB`)WN}Q5{#fCNudmo@>w3bQy4z4O(>lHi2JGm++hC z@hD2?^>%XF3fZ5fXME;q=*$lY@3E9rzmi9_Cf(LLo{ZP{$=!Axx4zQ-DIG%o#)m#Jb5CIC|T#5ixP#!2LW;8 zPu|QAVly}P2t`>f_D^xqGWvgYV(0&K;=8ItJ&h{u$yFl$A6eb{Dn7ZGqwc6)FS&d|?R0JQcOfsSjb0yDpsZD* zJ21TFw{B6YMZVerswY_UXhbJHR^n!c8OlepETioQt<$^z`)BIW(Cz`$@59u%9X%g0 zv>0RV8$NuZG(&DkJajFHTf{N&SSkfYbVqHY+4gWj-1X?(4U2(2VKKU|LVw!Qg|h*6OR62fFh^98lA_@G1a_1`@nEBTzL(F5j#)`T7%vPDB zLX?og;_89^FW*MV5<_-prS)S_M6~!jEI#4iwLxmCDcJPK32GtqNMVV{3!eQmdA+;C z*}ykZQYqMwfqVW@Yh1!-kYs}$U@1j0HJ3)mm>zd0ahuYFN%xH`B-2KJ2goI*l5Zmy z6NeTop}RJ8?iV7^rl~zM(Uu4kPip*ItNf?jftl`_N}D{{agm~k2TOAB4D(`Hk`&Km zpt6I0Cca=CD*CU|@Z6bY=otq8e;(~iFKCq`l|6n?vnUh?rSq5@&*ZJEKCrT%jAHlY zmlX~V8=r*zsRp2uFh})|+MwZV8U#4Hi!RAK?^ADCSeHO7*IXs-^T4LEsX`Mm&Mi7# zCD#OljsHTVHbn|nh_47u0+S%0eixCdW|Bn=<^Wzj9c!;Oxn)x%Pi<3HfKQ(tLZ{i4Ae;5Av6;te z2+ECz@dp6=hgj9eJjqwI>$s^o4NRLe%G)50-Yw1%)-w7~0VhG1AYF2r9nb%^9nJ&7 z1g?KFXk3}7b?pkKzKSRrG&wyBt2aulu)djNKsCMs*U=l5bZJxsY$7hHOMuaPhXN)N zt-F?*kZ+6%E~4#8v$?YKMq;`G?#k2pwex3p_-RtTvF_V{v`&vig>tWAZs~?VbYkV; zFg6|)e1(Q1wXCkPbUy)t%QC6qLL&E;>6Y6v-JE-4co;U@F?u zc5I;?F7>+{H65BnOrY%W$N|CIrhZqg%1O;U@I+|a!J_)+)^}X%M6^)N**Qj1@cexU2TBn;E1v@y?C7*o){<3Gs*_Egh}gqR>LhkH zMrBHUMo5NWIB&09I1zo+j+<&5f_P$sY`mx^fCBr$s3`9uPKY38(SsC1 zrT?%K;seQRm<<;}==R>|U3yxtrT7OXUeyz{spYjo zqaOU6S;j@C@ayh>c!(j3iJ%x>$bt2jV_OXFLnk4AsV^C|IUe@B)RBnXFQ#p`!EODb zkP_-tFY_1p?EU}Ul`*PD3zhs+v*fK4YJ*fpMqW;=SiL-d_`#*&W}ir&zxwb>chc}) zrsa9gywP)|vU&EG+gbAkRJSEu8j6lJFwQWI@IDX$E)~Ur$ctPIrqVA7s4`+@d!w98TdkV zunV_y|0d(9$7<^oWKJ_ZuE_XFN~i^+Kx;CqG7y^62VEWg53zbDHO)Wm&vgS*)>=uEc9 zgd2A;BK?^gN!E`cK|Z6}iDWu6MPUPXuFJ&$Jn!0n+>wgV4s0KuN)Su+YyM)dgC&=w z+)-E0_9W`ZP5ew{Du3=f)!Q*OvL?%)zMljyf7G0^I$G~K60cSMaeYQ7I7Xa}xhm>> z=?o{mrfL$(#J{k8&LCr`xO`j?Wp^{7P5|>EqNn%zsj#kuAZuGs?pk zeCKD+QMDh|AuCc2ll$p3%djE9PbGOc*;{R}AK#V;7%2t^C0}W^|7QrT1LYR3ku@w5 zZ&bakdSWw#KTi$1m>}p<>2TvU!VtK?MQjDk(_sV7O&DzG8q5 z*90cBPfVaN{t?#F-0f!HT#zo$@@K@`e?JS_pd+Zx7`c8k{Yo!tYRX^m3tDrkppRDq zf`;hlwND2?>QxgY5jyI)|2AE*b2s0So*rkifuF3;*<{xx(t_ADXRHZPnF$Xm=Y@i? zjeOmG{>K|A0(Hm*`42fH>QGl^7Lr%w7V|rpM?7Q~v)CJP80FMxDzAHh?QH|Uw@hW9 z#-0$=F{+V#&3DVb)>D*;`)oR&8vJUgIt7=zv9KJjj*P2(!HJ196$&??8IhkKO;dAn zBC!dh+puhcTT&5Ps$%7^z+8|Erb?9*D}%;tfW8L@QNsV7eRExH@SQ|-E)1z+&z_7i zO+A}%hT-Nd?T^P0h))t{1n{*6^Ga&}zPiStVAQ37#gpG<1V7jF0n~DgSuTyveDOAa z*rc5TuY%$geCjU$wa0Vm@*k6Xdz^g-6PYVMvr~VwtkSjI+5gg7(osa5zs>o?wF#OA zi&}4k-qRSPK4>rghM@cDKIax2K4^-V`IMnj<+6|vpkyHMsr{yvJL9Fgn4z6))1e@( zg+K(~V+?gfsN6`n2AoqV_4~eQ>b4W#urqxYACy(>B|)k5IlgwTRcy>GR%@Y{K?;oQ z8SVP;)+BkHeikZ`PfhzjR5rE{u0B^i0sc2^2loL9p%E~aKX%Dy1OKUiY^E+U1RhTP>KK=C z?w!1eaP6o7{kzBJ4HZ|Hs1D1sTIAC={sPsQ8G|{>tF}nTN=bfPeSDEFa~C`~az3aK zf=i0gF2y3>5ZFy*+tgu6?io)-SAZjy>@EjAR##emw_rq3@2nBNE1XCz6fgaoB1Bv& z))JMkV23{%^`=jli!^*|vshi^NwuU;$VV6I443&Q)WzyAlNqS1wDZ-KsQ!+1hdtN) zd>xgDL8fEOzLylz2Ot(pL>+O2kE5~Xk6Ehj7v0QEh1$lP%wKDEYDrUGg-`^P)(IXF!d(_U) zPvVJJec#0ToU)vf9=}KN8>U#>=QWQdNhcD_a6g~HsWVhh{UNupV5`L6!%MS4tkU`q z6E|42)8AMGJZ6Z~?WkCCYAsw!Iy%*ExJ%bdF-^tOkr;Bf%N0v(LpW-_yNkc+hk8CteVDj!|7S1#gFW`}|@#8&MZ; z5(gM~cYBGqxdaL9LxhF-ru#YE>Li0r!vTUX+FzdguKPc@SBpyaOuyLTx5Z2$trvQ% zBBWoeV?A*g>I`ns))XLRS~p zglL~s>cY1MG`hKWtV&Y3G^9K|UER_#-?l^qG)5$Iy_J`Zm)O?M~ zoelbq6IV{t&WRY02VS(h=e(*7!3&_lS$!%@fz}<&LMe!Ap_bnucsWi}MnGFOTlU0^ z#dtg4qq=O)I?*oz9x(`EDG1}M8 zqW`bV*!cb;m5kpWXF|6C~ z7hCf=rofPD-+T(!*>8ax-NndaoyQfaR z^1nEtlzctY#fy^jKjeDI&hP%1ZI@_$#!c1MwA3`3+P+UbO+Rdg{At1gw;e)P-|U6q z67!Yd1rN*7bLcB$9g=k$fA~KMWV76N?weO(v1DB82!SLz774D|EXKG@3Zr+?8O`c^ zhaNgxW3HaB$RTqtJsq;YwHgyxZ>+=;IGj5TiS_Z&(WE}F>c}zc{nEkR{B>88-6hVK zk^_Y&63?_Qe8E$~#h-TZmpK3CU-AVAOQ1~rZw;eZSnQd`lFj8`8$M#AqvnogFLZ@| z@n(aY?_jSyPv%_X%Wbla8UOuRm2RHN-y5dEN}t+o3EZID!|$Nw!!Seo1PzvsCco(I zzRUNYAXxtIo_5CFF)!3un~G7<^Bj-2Z%Y4e41A{Zj!M-+M!I{@=CgT`awdb&+&>-u z!|AgdhJ|`H?HKdll|jw0fhhEzTO#wh{k)h zIZi(+ex{&fRl`-bq4KM5@^0A3sBl&8!)#W-7|-{W$C!)(l1&QMbta_HQ7s33awNKB z1-PyS1RgAM1nR6kl@2xh(&jyt1}$qA**s&Avdz6-4fPAutM4yK4wIFZqnFE@jr=ZJ zJBjcdw-=}u6Xe>ikVvXb%L2H+*hNAjRR6@hP&oWkrsXJ>^n4g?tD_yL^lz(PEp7x4 zwYU-bIG42FnxTZBnMBZUJ`WQqrM+l>To=5FaQ;v^i5N64H~syK2H!jJX`#_2yJ5q_ zJ4(!lFh-tDSOt}f47+*%naECKZk@?7mv^FWeLnn%2>sV7m6X|_-ZQ)sf&jTrFsX!9 zCjK*l)`k2vxwEH?Q_S#&z6(E3a@{>h^`ty~^i5>T^TY^**V%qi!5%AfI9pFya-EM# zY}xHKMV^M5l0fUj2zT^?ay?5o*45u9QD@Vq`? z`B(pG`{e{>uXi6hZxDEbmcm2TweG$S9G$XR^*K)Es%X&iyMB)rh8#bY_S=7U$Ff9) zlrBKbWrBG4{8Dqe4HXVx3%6_F4^E;ks^8`0_rtQrsP;?Gh(G66hIb-5@4Zg)>v2qR z{5~;M^Y!q)Q;^#J5;5>!`%C<0B^xogRR~kt%Wg!VUz*@$FHX})`@9tv2>eI&@vi$2Qmut98BJ-_3 z8$T)pCkw<}XqWR=8N4<`@B=#thi(2g*kgqp+fC1OJ8q=#AGxpk1*tOMg!Y7@qRE&g z4XWDE55-3SCw^LVw2J!949GWr@ny>|B42d`?}s=o>a)Q(2K0K$o%8dhI7HFwt{=b8Ex zGu}27R_K@TTetf9J8w_UesdGoef^Pyx2g22m`0c5`&30=#;}lChwVJd$$eV8MBqDT zGuoZcV5ftLuU1kxtYI>?u2vM0b^sd8Yx5&HyCHRic1p{wdY8$jqVA9a!u#lfuE{fi z9rmbec*Hh6^zij9&=X5IO)JjvP7t+#=dckk`?{Cc;7u*c^XU9-y;-EWbkI5ZdFu%& z`@>(P2wFI@e+?|D;i1Bsi7hkzC^s5aso=KW*565+B((r{_sshet77#zlajh>EbDcj zL#(`xm1S(*2SNb--#)kxLsb0y*>uj@A2S@MjqZz0TZ^9pHPp={onOno9(Bk%%+_#> zC|F|^jsI|&mS;KTR;%rB=1{DK9^~7*W7qQwH-I8Hj7K70qfH019T1e-n$jW zt9kqysNjA)%wwE|!Wjz^RwJJJ;FKr0uNGu(rk~6DL7^l`%ei1CM8Y8vz2+SE7?>hv zQLt)cP!8*w789|g^10Pc*&Vn0Wj8O`Mf)UBtsLayyW!ek-Ri;VPJ5KKB;lZ=Lz}O` zlD7b3xDc=XS#IV-+D#U-+d}>nLv*t%bgsEHM8Wx!$1DaNb;%yQ17&q&BK@NB)lOd8 zX^OWF&l`DyyJh8l@8Ir5o=*>AMo{wJ`jB z-MyfnEnYeLS^#&gS<}90!^0r#bteJ11(C8ou||H!X#dO{yd%4Tmh0 zNzoUp>6RgY>jS1hv`kLBX~4%r0egJCGXlnLv5EF=9fy#dmp)ti(l;IzE%)V3z$fk9 z`v8~}o#;B54z0eEsMcLq6fD4v_x80nq7{oX<+n9(?PLD*_V{nFcjsB~Yom-04Azm| zVYcfkYN}wOkTx2j!nS}!0TI@&s*RtLW<|G0w|Zl}<95Du77YafuwQ&W6`$XHIo_TM zySuWq>!3PeVLQR-SjPR#*@0gj2x-#Bl|CZ?BblPlxcT>kDxOO%8;%%8)kn2X{}Ie~ zmTFsH{#L2-;9LINqs0p^Y1Ph`hb+LqD2wOMIZh^DiNk1K3nM>8xh_f6{V+<`nvURO z@Y!fQImS>S2N{bm2Q1iemUmdMd$oj{vX|VJQJ{~tFiITbuQIEU`7{q!l)EkEyCv_9 z!rlk)^#`u36ruD~6f`rK1(3WY0^SMC5Z;f@On)Cn7(KOTVvxZV*#1j;Q{cJF-ibv+ zO22LlQD5IsgN~elNa)^pG|~uR)y0jvG`e{_IPicW-6bi`hS+Y~yf1WCKMq@#hN zas=?nPX9x{mf6v>SG(d0J2MdDQkOreN{6HGL2qIudEGkvdNW!i{=&bwWXhwI)H;qq zfe8}bPJEdkWf)ZBiErA}%C~GR(L0@f3(DT(Hz}I-_nIQ|Bl}I=vX-LxY)j}x7k+Va z_Q!E3@fyjIa}%9W7mGm&o%yJQ)R&071P8m06+7Q?$asfJd!a~%E#>%>~)&kpP_|xt3vg*3un)?|+sOz2l*4E2i zo##sE)sUm{XrlCHK?OWi0%lsq+j5ld5-H*KJg626NufmhmAJ#dQvSZicZD#L*AKYu zXxh){)q^p4Zld|84A~Kl=!a{>eGkzVU1IRdZ6O(p--xoQ?V0jroGZ_B5?9Pm@yG_v z=acM!m;rW>2n%saqIg|<^XI#Yfwra=MWMm(<3xL0^K7@jr-jD<5t3hX{Yk0_2&af( z8SssvK=Yk!JH@l>H8xK=s)X|1J({|-{D2yPzc)^uyE<5HlX^{Q#-pA40W&YifciEg zu@cUHotvM(@FGx+7eh-sFJE(*I@9 zcGXZo=B;fXa3{=0l`*~k5naVYBm=chc#6p^hTVA|6o=b*8zp@GKBr`6VXn()EfNvr z#!$jI3iNSr+@3xtdOc^7Vs1qnrZ?zqJ-dCH$nA`ZqUnNuWAB9oi2MRfZmjg0{Fg-Ny&%oc!0b9KkQQ7ZazTQB+Cf#F&*U-q z%im>QtEqp>qDXTKi#fmJMU`Da+aGo_gqL1?yWw>AIW5CEym)jrHodL+mx7mq)(u=b zWv;W8#@PdBm>Ejl0+CN{d9B61?YR$_Caw{uPtF|DO_JT_M_Zd+Z&@S@-tc_GC!&7+ zBT{WWv*f^45oTvUQ{z6v5CNQr#);C4tnuqL*|8%V74s6pUqbYBb(@48Y99r^X~75| z37zy&m}e)P>>(z9ZH)o$Mrrw=8zD{Djga$?XksoYCqE!DAFYi-Fn7xvJSh zMM@%!9;R6%<$l2xhXZI!i}m#e#-z<4p((E)lKifZe+e5z1s-C_Y0&4*`S7JtqW1rk zaevpEWN~|L9~!n=ld_W^BLHhGq001!egg z7E0)J7X1Ckh)?2VmR=oK*KS6_6X_G9t1q1apaC zNWt4qu6Kq?ciR|{BKiyyJ-ZwioJHZE)&*}}OTSJ|S?0OO4KMiBpFG))T0`I7dlbT3 z6L&~>MiDp30=Tb~N|ryr8-Y}mggtf7Nh7&1L@*-g#a|J0sv#&1OEK*7A?Eoo@Fx&pn2@a>;P zq`r(`CE>XTR{a<7ywX4x^DDuF#ZM~MXvd76Zw8r^G=plw$9v+4@QmhPl>xilBLBSa zPfX(PQc~Va7sP86NY+76AqLt&4d+}EL3D6Zqaez?6RQToNN2wKkS1;Z zZ29bk^H+S;s~%}OcI`D0ZrGn@kB^^6{(mk2Z~J(*FZHt`oLIRh!3%O9GmoX@w(2(@{HqK9)%)sc)nWYVO>yhg`{r(t8?c$9uXPu&wi0D-qsNqUOV^=-ZOsASxRB> zL?}|>+qu!SHhX>uG{2jHFwHFs#Kl(dlYqyFAD&;HrK)qL@RN96+f6+MwK@O~+}OlR zgD`f(^OK}xg0a?wx9LU5PPa<66FH1067yKO4g64sx$B@ZyE7NL$@T?p%wt$%f zJG{HXoP$zrMF)a|=m?a(svJTz2Kv2Y6!)R&j48hn?NetZxr(Iw(P+UpCgHmzF+_X{ zO!_RAjtgs5h3OkzrzTRq!5~LhOoKCKtbkxQ@`ltZOMp&~h$)beOb?EDBD=;J>RVHcz z8ga&#OM7w%Wk-Xf@g0k-r0Hju(aRZt&G}&clj%2TG!zV!udRI)tLSRjg}N3!g;{*? zqCN2_2hNaWGlcUP9!Hi_%(4qa+~H_{a@#M`Cuf{wAi)xyPKdEhgY9ONPu9dy9oo#>)8jC2 zxFlS77@Q(dROeWdz=~?_a8L#(X_egm3$0uF_;UCQDg9x<&4gCOQa&gkx)QwDj!Ur~B>3`j81kK;8ZaMPPGcP@#~ z(mOC4fCwsGDU-TuKGcg{%00Z+LmcbMXm}G%Yt4P(q7%HUU<+$b7SNM$-}uWRh3FY_ zDZ>AF5UfC!pdh3(F{K>zYl8~aI@Rc9YK3_d_zk_Kohc}gw%}t_7Se#KM~68%OGErd z?c4yg;}GlMZYo33^K%>!?5KFT~4SQEdN*pi&a$=v3v$ zFE&nLzc4D&T5@7j>JwYRGA?d9KC<)#lTt-JAl-Hrw?_gQdm9{p&L<77TCayajU*NO z4t6I_HGDM+B?&1gYGe%VbF=6qls_B>EBh|9a=qcOHpe^nHjrQt6v`|5aQR`gJJU_l z&ez{^ASN;@qL;E4edzF&gDSC~9@aK+5N~D%i()jSEGx_&Lniy|`fL4`gxu6GZ^bo9 z2H6KB4)KlPy21Tip-9uhd`tv!r^d~gfst;V{!ukz^XO~NuNMyr5K;x;^ zjF$VAQKp^b*?)GUu^cVV;*<#qum}Ch>zc`kXaza|jqR?vK&kInsKBelCW^JW74HPR z(Js(o&Sjtsvb21`?eu0FrS3q`AM9P9$O48a!MxZ!w)fve8?7RAdctF z{neH0Kq1jdX3W46*7oM}4}H{905PmM*O2=!#L2&xAe6OWD%0!6F83sID-A#ppT34O`LG|Z_b zS*eHw&G`PfJ+I$o$Ul?&U0~PO_kZV1?|B6fFdCtGw(9=DDbR6Wjy6nt;>4k1CT{FL zYNq{6~2WEK^SVB#v4 zVn$SfK6Cx)Fg210`gHOh;Hkk9LDns@Rz+V6_?La6*EQ!Ig`0PA>a!`)KN}!EGd%{{ zf80%NsL>6|{T^#h1D$#AUjkRSod+=riYv z|8@A+q%1A!MO#wib))WwB3`dl;;F#%U_=LjmiG7QP&ZhEFXP4@-T2+fhV8LO;ND3l8lmHr%Jcr} zp7PqwC$8q#o(GAHXNF=6y{qK-X1EA#jn>~5OfU6i>r-%I;mU!W!bOAf8e!7Te|?1$ z1sGYwsYSx}2&W$WKDXn->GXDxgM7kD53YJ>EFzY{#n+oM$90C^OC+Tqis2Hr(o| z*U(m^V>cF7W(0j98|^6Fm!Firu~9kXdVdMMw1OD|5U4BwIWB^GlJ2=lXk~RXP=uXxqz&_5%po~!tmr%v4vy;Go1M)bh9N4ackJnqJkbA2lvlPFz7Qd1p zD~{u*cxrIZ&aG<;gRYOWQnZB|%>jCuS4^b)i;hu)qwzMw%OrteEHXBFg1Z-;0!V5Tx z6d*H%L8-#{;+7jmchq;Z$hJSoy5Fg&#M4T>imL}BA)ShUW5##B%~u>D!7kcvoWdfL zB%f{p<8HSa)|b}kKoozh+jdcEvkkb!etj(D+z9c+X5$JbC{d)bkLd1ki z94E-HKu)B7yB|E5!CZHKT^14KLBY4{z=<9jKU~}P{ML`6x+i)o@lp!!eMmuiYo|%z zK6F;#5cmV4WGHD3GDy$mGSx#1AkXGcfyTzhXFHv*8HHVx0uUdn{Y~JX4w!(|h{ei+ zD*!5Q+Oi!+Oly|0`Sheed24lcmdj_pcn=hKQ-pOS7BC_yxA?btlD~k6%#TEL#P|%* zBT6X@f+UVQPCDWg=8XF``i>k?E~XaAcr*QkSA|pF;~139_FyeqJm2L+-!;1C_GB;> zA|^LXjv8S{esYDmbzOxrMTws{xuH7=j}Jwn>HBs-Nf|PqDLoW(UCctL16H2=+lFK# z0sMwfy{okBtcqoNg|4ddZT#7jEeMx88o8^%3s`gZiW2hEDMWm ze}gagi=DQ1ANP^b%1NoWJ7j461Z>G=2biJ$+s`{OZNNv!aqJ5SHPIDY5NqFhBKs97Ckx+scWuo~?nN>T89 z0Foh(x={Oe_M#hhbbq6g%QIOZ8}Uq1;QnB#XP}LK=L`b53N-{V8RPq@n{!`fR&rNH z{#AUgE@<`1u#k%+2#zEm;MHp)-~(HwQZC>;B<;v3wnwWN5%Ee#f=0*I8ilTJJ4ds@ zS^&i3i03-59fi7YMhD8x#gM(ISH@BMqWri`#Ow$Yi?m;o9^LGaCJcXu3WqC64*9F)i`>rlM$~IJ7Nf?f7s3q<>O7A>~%LMW%Actk*XbBq1(FR*NC!gwTNIUAa=x|J$hmRwJHy0N? zx#cWN%B%d%i4psZnDt-5Iz_xcoJLK8;8e8hY>Q;Pws&vVf^MQkR{OI&rm+`YNvh{Q zoW%n?RMefECU=?7VWHI@-y|bqcApvTe_~U4Y3nPvnv;V1=Gp26CDm17N4|!$PCOde zl(}h1RrXGr)K8f{6mksmL3q>xQlT*mYVoaPG@QU4r@R$m5;HLyA#+} zgB;;SMFhSmCtxa(BVW}2g0*Ui(H0)I60ZgXJ@h}ULnt6eWy}s9Jwt{GBdMQM_C+dU}e|KSXAjv%co%*a5>p4dxDiT zFA5n(ske!g?Qd@Y3VBH{dRZE;bR2YcPJZ1griMS0|eb*)-QARle5zy9YfQJWJLg>-ysy-Q>R3GsRH3e%l0dt$8t zx*&aPRK17xK?iCtx2~et_TO&4@CGOQNCrTHDy4hKL04;y&Rtkf&s|u%6L03?wGUyF zfATqfSG$HA(N8SI%yUOEE)U~xcI`F{nj5H{_Eq8F0>iW zFgdpkP$*c5Ef2>}??ZI)9uH-(1;&e>Qv@WmZ~~2>yau299E8u^W%DfvwFR;;LXGEC zF6+&uES7fJJz~=9jKwSU1fi^&=w%+0S9_COWwhChearTk2hcwx&~9aXzzLh0tn3*Z z(b&^_ll&!CV?N54M6__QlbK-aGhy?kE0H=tIeTFnJPL06Fm|ni4{{A_!l4LyFtjQx z$h!=DL4hLUs7_=5V+6`7v&7JmH{;l4+Nz*C(DDGh`T%SuI0MW}p8=_Q`E@GyISt4^ zZF9%0BSrEfI#0z`2`%X8rj`EkQ!;^-C;avOGDU`@g0g=1e}vO^(_l?MhXDgbJfA~a z8gkO(F1MKqMO9x@x@kDtM4Mn`)Ac4lHJ?2x5I3Lf;Cdl0d(10tZImkIr^c z4WbB=cA-q{lX#;gvS{8!xD0r&Y4>wJfb}1A)}hU>uV#$2?pItKFEW>K<4t7HR$6*k zv&UUsg?$?{mZe}D) zR%!PH+mtg;JGxOm3hhA`oI9C6AM z{9~xwS}VfqNeT&(!o%MTOI3IMAT=2+cNB$ z$NPM5D`WVFkyZeSW6_>En~pE>qdLB^$HkYoNqFF^fLI=MqpRozX}qmfOW}o4gC_Q` zS2-6L&k?eGKas`*oQAROcG;B*k$Y^N$#cbvCh3irzS-Sn)}6n2bCfe$e2@w`!8x})sjY_PI8T`^v^qJ98!`p~$II~$Fs+?wih=h00-XuG{7 z{6a)O+fMq~S69Dc$S5oAq*hVon)ZyAK1DW;!bIBpeLuN@biWI9Prsh36MP1r%p%mg zW?{O=q#q?spB~R0ed$W~_w~dkdlMn?hhU_J;JJeVDsPT}29-Jch~w8b?sm_`&`@F$ zS@hOBj#Kh(MR75M@0EfAM`-_iK=h5=lJ1zb)IJNp;glSM7Wd*6r?Dz^B9+&BW3oeH z|1_DKtk544+D8Xr%v0EJZLHYI7OgfIS$yq1QG(Fd*A=8cimDjB-(zPy1N16g_F~K6 zHTFRE`eqMp=Q&v>h8FqQ^g$7CF&Cc^pNE(qW-C-=I9bQ*fJXg(hb{)+Y=i` zeK@>jNI!N`bO<@nI8F_&@8Q0!N0m1+psW9Y(kYmb3svY}bPbr62`gSMs=B-YC03?xx`^*+PBMXEXzgc!;};H)6rZuq)hjuRe$@c7^gvsgSNl{3SpfMD!f*5% zzov?65JJBzkq)Ew{#?JE@R4>|(tQ`WhDc+OL~#P|d!3h6b6M$EB!FB z%CN$e#0h6xG=eKY%5y$=M`D}R2F3KoSomUcWKi9(XUhLKGiFA30eEpF#r(0m&Tsy( z@Y+>;zL;AuAL#Kxfh$o0zkm%R9ojon0-ZhRvp=9EQi8plMesD5EeeXdb6*Aw9;%t& zk^VXGt86H>Nw@%=Em|+0um(!&8yN73@LoF%{K=er;kd(wTDn>bv|8-*(G zyc`=+x2Za`U!qtkUHM~U{re4SBe3qTJSiWao*jx!Twa=!E+Z8>KYT~@L2(6jcuHN~ zXt%s#ZTP$Q?!tz+Y$yq0j_nX+&-O%yt49_Pesmt5J;93{cDQ{XY1KKzR(}3rFhRg? zlqNcJcd+uHmeQHmHgxt*e4K`7(H`454a)m#iJGf}sR^!zk%KNV@ov!@44fyMC^Xm7tI5tP+F;dd@y_b<;@yG9vQs^TeJXH{`wITwLkfJ zLOy)gNNN$m_(oZGEEUK_^CPV3|9!k*L$icw%L{a3L{Tj#mG0kMpFu7R=GGpa{Paax zRsgnKFE0EfsBj?`NC=TuNqvcu@XukiNsF(hd~oEL6pi>HsMlw7XY$6V)(*v$2?2gy zWY|WIhK>$_Iqcucn> z5#K1|C;JEMGSA2M?tH;r=Xa=V#5vASc1BtwV$8&9gfxN*MflIaGYL; zLtl?GAgq*yg<(1a`t9IA;1JBbGc^mUb{bUpac`D2L6@(Y zN$Ze^!%EX4r2mVC%AJNG*AA@pOnoCXm|k}ylhK}FCsFfT9EeaR?SshPkt0qoc#9UL zfzqLkzUd+=Ew?gGh{h4K9{}a(9svEQSQN#*LE?hKncTT4xJ~Is%lVcXCL%0sJvAB5RVC0pf>=trUDAd-(in4N|$+fAlBI4R*==VD+7Jj>h_l)T7eR5dNu#K^f@hLq(-LV5@V0%P^IDKxl<^9R+xZ+7tylW+iCo{K7n3( ztJu5u(!00Y|3WUmABuFsB@JvHpljA$j(Lb%-aCoAOMy!WY-zzueL5xbdodcn1?&Ff z8cg$=1QG7VlJD54FK8fjX}`t#LHf=VbryU7RAzXRGphC%ia#(0y&4Eaed%rqwa-r% zXG+E@wHk(T0Rqw+*YHaeao)j6^Q8MNb4KRpXRNSB{MS&=EhB|3jzgzT2H=$V)7H+h zz?=iCR%B%Wy~|0RNe=z$Ybfsr8IMX(;H*~#0ki)IZE+J0u6--Yt{$a>yEz$ZyFXF7 zD&okl|3X0>|3F|4tQDeD*%f@hB5#cLMR`J;SxJAKJju32Q`j}3NF9yCe1gs%5C zr{SW);N35`mg^Wlw^tHM+NNs%$FLl~Uj6`tI58E}yxyh+Op(#ISzSBVo^Jt1>s9U} z$)BJ`uZuL~J>weew#9TH=z9~g)eC06`2I}kV%z){;tIZRJMRw{1{IooQ-~5Ix55c5{bZdb{)rv$PQa&l0PJt;kRnX9gk=?jBV^t~#j{(3 zA3b+V-BVOU&Wyz&Nhf5RWybt7x4f{m|b~pf%#@eQ+hwU}!1%{`Y3mM{AGE+>CKsZ9;ortb^)W$}ZdNntd zzLpE+5RX2THyhI5JDcxk(Mwuj@EX4<(oI| zXeK6VN>xGVWb6mL6jXh}+|LIWY-~Ya(A?g$a2n1pXdRW6vAtE@2A{5q7l1^;qu&W& zS*CH@39th!unl74-~8s!_HPdv-YBhjO^Jb%K?=oQnZvj(g+MLxfM;jeqt|P^{^8gv z*P}~o=@eD0Csai7KK9eXc=8RSZTWv*kR4UGkoWcz`|M6w%BmCWSR?!nklKGL0ZIpO z<=6VrFRfnQ7d={C-Nrgr?dNa-*Gd-12rvFW$m47r$N$6LTSdhcblsw8&;$*T;10nh z0fKaJf?JRTXk3H4c5tV04G`Reh2S*aSb!kG9fG?<%jtao_{Ket_wn9w&KtdZk4=r; zt5#LbIoDjV=TSBO4EbuJGO{Vx=x@V_4E9>)`Nc)3tM-M*^CTm^e{g1!Y;}@?NbKT} zpGDfFX{ZNe>Jh#hFqIkgGP1){`@b;XZSs{0dI(d#+{J^&R&GkeUks8p0IfZN-rySL zCpW~n_=(c+VtyP21*PuH-Rd5;!`Z}PqoYfE1jA2|n;zM!bl-~Hat}Er-uXccW;DBSw6es;6GU5C-!z(8XU;VU)phvl6MS~gWR~(icih_8B#0Hk zqL%`o61lJ94n`V(I-Oz?8twODXy^+rY?s(eLF&-}vMzyOZ0eucyoXLhNQtXHO4r3Z zA@IWJ&h>45WB{DDe+jVbyi+|2GK34k+x?-MM(ft7Ru^wU36j2W!-%JMUrd!8KYjoZ zdN*$$PxyTu#J-(D`n#5?s3FWqWQD0w74}8XqKveVTItvwk&`oG!oUJQEL6oV=HDtH{NVZhqn5FRa!N;S1ajBaPMp%> z?iL4hhI4@9V28Y<57304Qz8wZ)spwyWUtD+@~VT8sGh-*@_(F_ZCW#X<%^eV!U0R5 z*YGO6&q!5-*tA4IYyy(%P4S;nbV6k^?OzzeECyqP+H*r;hm2+WzSD_+ElJ6s&(~1S z*?@|O)B(>S;j%sY9T&+5ddLrnyDMpz!)c-ir{*eu;X8}#y>km(4c$*c-YTB3dDo|M zIKE$X%lbA8xIT&hUTqN5KSVYGUJ3n_>5`L}E`M`QTK7Xzy1$4nWO?z-&U+fZP7>j#< zpcL@Y8Mi1utw<@W6oc`X28f;ayno&XPkD6xSq7ONPok5iUH>=pSrpCKlq&*Wh~e>)AoT6&3upJ)8~HTnyl5ZVE6&;72M z6-mGk74SV4&YEC9hY!&++H6fkT<3of6RqdJan>}U;p+oGGcf7$J`2U|0#BYZW4+x4 zbUD*l9g^`vKmRpbEuv~vJbBS$KFAs932`WIed(SLzIQ$`-SD5}w-q|gtC#GJ_s|;GTK4YMr#`6@bKPvY`1NaC>4sTU zD)RKDWv?-#2QSx!Cjui3Csr;afw}LyF6*Die2o#EarNoThev3hB+-XlbsVx|p%%Ho z=Iy3BpVe)Vw|Z){nWTKH{_6x9#phlIQRiyi0)?laJptJ{tZoU{J?vHblPf zun~wq1{b*P^4cv(KW3T*Ip9)UIjXK&fV6SFqDJvT#}}Q5Wl3|- zgZ&}VGG`^dqOXvWxqLB@Ob47w+cn7Wh zC3e0K`~o7Nloh#02}8&Vor(bc#V2+%_i8Yb^amh^+8`1DFQfd?7@E=wd>pMc6X{v1 zE=~C2_blZ*s<|({-o!==qkA!m9=Y8lb$!b#T%U(LP-zjR_1y2H7yEe*kQ6BsoG>%b z^L8FW&%XuN@s?>!RB-z~x;L*h_lDg9Vc1>=Fcf#M*zzgP)AD{e|HQdP=vz;IyQnBi z(<%A5ETn}f(?EI;=MXH9Y&{)J-U3k8bw%COz45JhOF*r04{Y}>`?<~5 z=O`%~iH928p$g>iC~|e`4|`r}!%uBo3b)?;WQVkE7)8qOKJtPF7Z=yg z&Iw`2qu@w##32Mx>h+VKatA@4RD}6DUS5H4t+Etw(qESJ1o+lYoi` z**ykIiN!n6Ik=1CX1SoRJ}Kd1w-3Vh3n zc)Kia0cUOs@o|h<=WuNP{#v~XZkK{YAz1NZtQNg%eup(wHl{qVxkAuP^c%!)L>+ho zIBk4R;@^dM_lbpjk>CDC=u~U@*G9(5D zK5MoiCT4#J%FY7ah}LVcp+7qmO)Ig+XkXr>)e8&kW9!`PCjI@+<^fFIv>#m%S=V!B;gQ+H4XJyJo{GUAc}o^2 zbNBw-C^S+4r|(wy-|Kwaz;3u~aU!m_0@lPf=a#+rKl&9S zz0*6mbng&uB{(bpf`@)5GWt|*>9QMIGT)g}roIn{u3B(w=t*En#s}e(<#!O#O3?hK z9l;9dV7ePl0VJIL?0zKh@jmoA*de&Ecx>(j<@((hpLriIro~`wj)Uf$QdqqNY$~F~ zm-4M_KKalQ^c%<|P-QUI!mr&zB1S>x0UfuD*NknWiP9zsxaHWH`KS3azjQHW&DwV! z6pcnKgAl4j?TP*Vf8(>1z$H+^kgM8NXC`HEV1&n=)BD}mUT_t^U0+@m_3!@4V(z{Q zw?ZG&r_oQuJO52nBfa~##{OkMb!h#Djda=G`_W6WK{vo(ErvlsIgO(H@WyY<{>cN9 zf4eMZ;kX}AS6Tk|t3kE95V7LMN%WggqLnUuo{hJ^UQTEX0e-6%4AuAK`;EJG-JVSN zL+*b$pWPmwMV>j6AYMAA8b8bFW~R}z6R*!P8}&SA3iky~p)5{)U=$V@=`hvUq}wR@eQ{vx<%&=q}{ zl<3hC7(lT-@U#BLm&Q&^3i_x!-nE)gK-=l<;P+_`uzPy#>;Qt=9jt+au9$zQCD)9Q z++7+URGdn)7ek~oWU_dGQ;-hWe004k@+)Mnss`x`=ykQKx`CiJ!uneIZe4_+{Z)Vc z8PLJEOXt7-cWZ@XM^#w}$S+3O zR%*%6aH^jPQO=PyzGnHRb11K2xOBNYvnU`j3zxw0AxE9^<&odtBe!k>gMOI-e=g`J zFa30^Iob=pL5PILzPOi1<2zki-e-G6_qYKC8M_`p%j%fHit_QZ;?7aIz4}Q;j=MKz zLBED*nW&-Aw%nw)){Bdci!*x7s@&y^K<2CDJUw@0|G^c+|xf? z(RIPZ5`2|RLRaxL0brc8P6~xhFqR9^X zDJFhJH5$(r>budqDuip<0x-cMvt?F42v*ZVrzqpNG;&V8^*W(?=y>9sOi@6urmwx9 zco6z*`-3+OKVMwpGj#7ep{-Zs)(vM|mV}@gbr)iHUIS_&;X6Zqn%6=fV0DW7?&ntk z2+3_8i2?cgmvucC@!gALJ(8~e{JX0%g%fvwTDjsT2zG}zX@Ef_?sgc|QAW5?9y=R^ zqyYp~zZC&}U%Xm)I;-C$7}0jiKE>Fd+k!y!UATMqad>H&ZsG`R<-=0pRJNDT2!d*c z9@ZjnBS~lu2T&vy*SC9wJL}`bE#M@9mX2h;M|M|g$cPC;pqN)=_Gw_Tq2p&5@ z^#MsMt}uhptJNin=zzD%HvjBnoB>gY13CP=V*d(ypwvyY14*2Lvs7NR!v*udZh#et z*m5Cn>T@{w0vZ-+`A&?bHHU^Hl(kD1L*a!eXtN%9>7d`$jPb=-vZ&DJgNxW^x$ci8 zQja9xq0zP1^-(1-1cCr7(2m!yyF^`>_I~2Llgw%{#+WT$ zkU?|!YlYu~8xuTJVL`mPI8zkKqLGr3O~RBtFd- z=}NA5r)RfdY}oihS77^9#6@XlX=!QIF#K_@8o7$arF54)oUh0}l>m8MZNF3Qun3QL zT>8^nP3&bLN7G-Rl${}GW(XW}97YHpKMzoYFLKouXn0maV55*9%=q#cqNqY;;1DzO zw2?vZ8?fdk<=+mAU~ke)oxdOM-J92LU3+d6?{XjKzyK3ZpL|R^89p&)jm|RE)^K2e zsB$9kB)`w;O8*oAlcWj0lwIySPY7bZR8PR^p7Tz5ea~lBz4I*n=L1T0jk&d8y#OF| z=en?g50JG#<27r>_SX75;1hV-37j(MtLAocyM9OkK-c0A=k#wk?n+8wcA#8L7O?C7H z0^9VVuZ%nHko&86bhYvynMfD&Jx%&0o}9k&&GGgzRV@*Icm$QTJ)Y3sJ%(&LUhLj! zz()_){vaH$MWUC!;~&tU#xpt}gf{(x$k4tC_D)n^-?=VBxXz(@D)5?5$ zU&;dq!*U)#A`jxoJN2md|$gsH+%-cm3`u4YPozqEob3 z>7^&2@U4t$Igb50Rdn!MHfZd)hzTNsYM zqTT9N9`@u*Yl70eknYsg(x>Ajc1cmE~;I!>*Ho3I?NyL^+OAJeColChke5)y!Tj@(BI%pfIw zb%|_p3BWl4=sh}^<#~#J;>RQawyrxb=bZTMQ#?1W2X-x`&@aTG6voI zS8TNmO;bpYZ`tneeIf2^i44tCNEUoMVmj%RLPFypRSvj@b~p3;f$itTdJpvfdEBY@ z2{q}q-b20|%-WxFwnv=MVlg}rnZrY4tvefNpGTIsLn(sRI-} z`48th0izrbNesSDTpUI?C?d7gB6YjVr; z{LdSP$&)tt>r6>WNOugeP|2(O>Nm-Nfm?c>2!kZuR81`Me&u&3obmP5r6#?hw_J?1 zh1w5x!+Ucf zTvIO#XhwLsba08>_?&W9SMX)H#KpZ=h2px*Fpp8@E#Q4AgU#V_VV_f%EGil8CK zVgY;n32?(|*Pla8yUvZ0dM}zRIAjUjdA$064xVe-`*&EHMPq=B%Dd^C1F&*is%PVg zW#`#;sfg6k)m^a>+dK$Axd26I*17a38iWxhbdRdz7HLC-1B8iA6!w8hpd{qt^~BSf zY6syq?e!z+=qHcVN;z>tsZ?V}z~YROq^UAdbX(IL zaE;2SE2xjjpI*^w7~+H8m+;P$bD`j3o4_AefJVxms2@H~`pG?79P+YsP$9o3>XHdc zoO%zyeNYLq9!&nYCG#!Z*8RV?ut82te0~)TKyyL~}CC&dSiFGImMxNk;qR z_ue<<2NjA5$399$a5AMCX6+c(M^Mcjbby#o-~8`&<1|61?W?rJXggV5ZaYHJrc7c> zE8omYdz&!{#dPuuas(T7260%KX%gr+WD4o#pH!{(M_;CJ!!X}WyB?6J#U_zqkgj`F z>roa5D5d_+NE{RLC&Mi;sU>7gdtN5O#y3;@fj%q~Z=xiJH4VW&%L$+G=zEPNqc!=U z{hg0gKhY5X)6f|=k&ZGllvrq^)uXX^3R(QN1exX367h}$|08hZT7DB>P{oYkMc$4X zO3*-TPE1EBGd?z9fzv02$Mw~|_-vuk#nUoQZctt-e7 zQZE%k`+g{SF1N`#&pJI*!`bdGokMsci52_oni^C`wHZzzf2$qtamvt?97#G?X5Kkj zj2WgB=_)7l6*Eih_UXC+H1>rI4iBs-tdJuHCY1HgHF<~(<)msy6s!!-8?mRrNhP0Lj5?>!rZOXikiT!wc z+&GYIN~~2vh*v+eT>=*}Q$n&2LaWr9WBJ(-eKccNSE<)eOE#q`!}rj07~&N&M5T}B zOc3k)1Y})I7}m>k8zq|eXpNxiSjv~2TC~KI--)VZI~>BqRsFuP$*o7>xV{7WI@2KI zsYP_bk~cL6UH+%~&$nWRXtj~lJjhNu$Z(IZiC&I7_OOoR0BD1|e4|V9Tcji_fDxBIQF9aSv$fh#-qJ zP>~xJO0~m!PkzsCs3d%@b;TEh5j=&oPw!_=jEl3b2Miz@GE&bK6_^)vPdTh$X{I5f z$pIsGIhKrg1Bx#|MI6K&Th70deXx>IIW;U8<^5y@6L3iW9K)rog!dbZ??z;psaa&Q zY1`DSP6*57>mB}%MYOMFx)_`JZ@F`Z+AmO( zo*ZJ)g$2yto`-eHA%anPXPAZb1tk~vOax}ARp~zyw@@1IP1$@tLlOh{6A~^9wY=|6 zA!-kF&04AY%&bA>tcr%gIDkVc+cUY9!2e9=g>t9C=E7 z9xq!kG3SF#A-YREGy=J8Ml~l85?&SOjD1@R1Yp0?P+)wG3!0eSkXm0P&H)@^U7wg`Rf)(f6u$DJ~LC!W0ahX{srClh%=kqps-vCp36-F&uyo&;yN#nbH{rBW{T0m4Rs?$2M;pa$$r1Qct~wE+nnv(^ zYUJ&=N{MgLUr0k$(b(g3&6RPjUEk^4d?mZg8r6Bu3sWnUM`#k;%4sw7!_y+9d&_0B zx;Qfj(A5^g#|U3%l)W;#s@+=MVDmqWWZy+`96h=fhYCEy(D>1r6_xXjrnO6 z5W>KavbDrUvW_ zT~ti1K~T^lxhMshk%d8Uhtf;$BgRa#i}#+lmy#S5LbQ>v&6-Km#d&(vc(tJaq%`=yBf~bL5 z>K{4RL;l_P>8O@P^B*XV%5~vU6iin{er${xWLoA1!G3fDcf@1-_Z=&P`6!p=4~@=a zx>uYjM))1NftI#ds2QsHg>>Lp4{1>U@mDB~l)@uX{M`vLr8V$%xO|xkB_E3(A^PW! z;dL0T+W;wK+5_P9njx;*mdCXTXY)@sjl#8Fu%QrNINPMC|9~3fkP6M;COGq0tb}UV z^myVIQwgPT+>j9=gmoDKk_lOe+J8KMbjN;FxvWt0wYItMNP_}-BElGDi=qc{^hnhy zH8>d(r-V2A8Kaf(f_pN7au#EH=d7Y6A&C30b`C(Olr zK;C%Nl1d~z>mglhZ!*PLu&(t0Bu@O^yAC-wzX^2gHk_YGIyb!scA?LHTNH+To zNO|{fH}syY+JXF07Ek!7@~Q2xS7xALg8or0-o&=f5nqWG=&ABs7Am>{1jQqlG5;)o zQw#b{+4Qk4VDK5qi9Ie)VFQ0h%MSz3_D_}j6_|qIe<0|#sg3i1n6Xs|yC;kaD-3DG zzfQ{?`3zv;T8Ql}`3ShRDwkNVDYaPOwb0H(PL{vfji~!lnuuT3Qy+lNa&lC6(@a?S z-mg$z3Gs-DYCiMeAQ=wpnW(EK{%@v>fuqWYdR4P~!e@))uQ{ey)P)GWx#A!n3C&nP z_nR`f8wBa+HWc$OZgU!T3VkL~v(@E@w^zo0my(z?6P0%MwB~Gr#lF{WC@uQ^so>EQ z6Vyp(M~3r(-Rc#5qQBdw9$;pgqI9Hn1bTrK^wgYGvrRV{q{o$YLuEAG*oeH;TOy^> z4iMJ0!O|loxO>A}&7u!}K%`22YWbG35r^@sbN{tADrN0{R;LWl%x- z8c@5isR>ESwpJkv1=EYrwkg92VkkSrHIP#8;jN)&u4V-X9&JopI%{#NY8raX^Vjd- z=?+&rsI>6P4EBPKu0lakk(G^ygTZ5OU77~qrDb>SYJ8lZE-pwG_rwFSnw}oEX&BNB z@goTuwKrs5XX5&L%~yEIJ^Y2DdaIFLtUUkTeizt^GqR}ivw z@X90jfB~uA0I2pXuf^(CbZhRts~RAFzVLBEd99jk(=Jiz`PB%!!RV&^JIVy+5IQx* zc>9fftWVqYR*Y$!5Bd+lnJ5WU+JRsA$M64my2+*FGh8i13vq#m+nV3?oB~nxi5Hj4>9@!R{ zgN!`#yWl0T-Cd-H>R+%SeZ*Oyo4?bI$LqK^;+_YolpGKC%+BQ>j=+M;gYdP(X8Ly7 zH;&8XJMQO2Dzh7XEvb1Y3N!RKs|J1SAYK!nQS*e~?3LBFPcDb`*GPIh8Gp7|6sa8j zc$eYJjA}BpoY2kqp(rZ&iP<1Nk_p5i^BaKQ)28_^&rXW=W;Jc6SV#tBsx*EL3wqcB zw6fTgwCceE;{+bQ@m#2{0I1J=UdeZBpmC);%MwPVQ$-1qKH&_T|3yPn7hWdJ{^>5q z)`(6^laaXV%5==qa2rUL#+4Z$6@zG1Qbuy!Ks^}^Qp`b&#KZmZs7$N+8g;Xl%(pmz z7So@Mk-JYi&qeqz?m4zA*^&&3uqi`nDT+BfFC3Fne`$ztk%d~4hBAKj1UwE`riSl{;b^@+f4p`&FcsOm?z z$R2^(U!E^OBA=IWg92wu2r_z|Bl`33rZi6b9jK5fDPQbBTBoRWBw0ms20SMxcc2%l zJ-hEE$oS#;b)+}R=&Rfj#K`g+O% z;C{LyoC>R_%jI0`n(vJnM@~?u01;Z4D-e%T@lOx<;eg(rEZFn6I#k(lqf4^y^;=f~+C5*9$Co2@ie34 z5e@&D0V=@C*`Bg=w^mbmxe!|8i5U$Cki%mOzX=)0R1?qiK5`}9Ajord;94H*T zYbfRT&$pl|fnKt7_-4NBJ0?=6#eZXanBJnf5BB;uL~JIMKuyKAibe9hkUXodlv6GO z0G6ja=>F-5X^^-?{0fUGT2@vz8%Of`q9W7D&tgKsJh89l15)D?lCwx-5y^I=W|}b+ zFuQ-HvYlEt5|0C(IdEDiG?w^N5ev(KJ8!PV%Qic@ri!!rsQy9cHkmd)y_plxTOZXc zWJcz#0*Kw@850{%T#O>4hPr>+-gUwQ)sx(lX=j)t6N*8c~o(|@3^ zk*^1EEdmXd{fEA1d#Y8W?+sGzi`+iu^F;U*SiFA8B~i^%!)#2ZzyAzN;)w+VTS@Ad zzdx2KF{SX4ZA!%M8?)$)E5Y&zEp<*}*{oFVUsn4;zab2$lgz=Z;7l$JTFr<2?af}! z8}@_u`2Ct}tSautZjkwWa}Z0gIq1+BL*P`x^IbheqyWAnGbAbc{V!8jUVi^(UzU)~ ztEHsLX*t0UMgz|reF@u{O{9GBQ?hB*roag&74WglyszH}Z{qAl`p(Iqq4Bt@FAo6W zcwN}4e@@YjFTRc5&t+oUa%7u7{4OGl`h!mY2|4~%oNEyi6LZwA(8Kdhlrb}+8u{J0 zz4H1imM7aulWuowAR#3ZsZjOM$Dd7;TiZ@edBXIk)Qup7P!nC89)NeATyf4g#u2ITUd?DYGP^RgF{3!!owzf?6izUlQHDRVr z1{Ikxp}BuQ<;ctOVqv0I4Xp57y)!_cr7_cU(kSvj&KO#w^EfR%+zWkgq{Kd0J-e5P z=cAzbDLvY!-h~RL$OtUYf9q@hQmiE-1KD2W5>2Up%Y1%BS6EQLuKNI@O)Bji>aRBS z?50={@$cXOapIxAnSo{tBKid~lHF9M-XQYCy}zSq?dn`2ccXi$wuJUD|=dXtr-8z<7w&&vP(=Akf*68K{W;Yxo~BUoyMP9--sr<}O$jk@QSCEQ{u{il;eym==@ zGVb_zcyx_AOkcDzY@RJ6fGwzcB^`EY=l6n}RI`!@-_5 zt$Rmqf2G1X*J4F6;$fd7!%LP@2=FiO+K@CSf8Vth2$_I*%Uzp?QQ=wd(pcbkpQbZ= zboAB7Z8tGv;2L~4Bkqhj@&#(BWnUGKY-|+b=v$ig9Rdr&@fW{98(nR1M~9eK`w2*P z?FjtSJS@@gIdb+g-;ZNfU9!&_AxuDJGLT_FDan&JCVPZ$Mnv;?F7H3Qx++wZ?vQz&H2M%De)i7|6;K*{&}DbcCE|A^I$Io54{TILR+@&w zT$`}`ME^0Wm833sh7&8QBMRFzPkou@ukYLmbQzZ%k#*8$#QKe+gyl4o9Mgf)FXDwe zv*tE%JAce}lF(O6B?dA3z$&iIn^SF3BS3}v?tf5MUnriWtjQ5;1!SO@>MTxw2lX)2 zcplb{=aDUFJ?`+8+<`FU`p@I28xPnlo9bJ4sv?9tX#(Ya1Mp17e)YL|{(hTfHMAoH z51HE%5a{EB!**bud-Xg}eG+_3Tjh1UT(eaW_p$*S0f7W+Qd&X;E}yVz6Pfl(7SYDI zkj7%RpFlT*HYAdqr6O=kDr}s^J)-9VHO>;2AAa>JiNv;87ofX`eq6v0% zl%BDg$2ivOloGsjpXrwqVhbay{#9otJM+F2m8570#V${ow`+`7OtDJv+$vc&V#)Ni zu+D;;!F`?4z@tezfsW{hlhctu8rjI}9*%58>_{Sg3h+mIy>@?FpF9{y!yso!lj=84 zi>9;8#Msu4$l1Bu=-urZmadb0{B3`ZUCB1C2t~hoqg@7SMD=QmYv8tuV8=4-@S6bH z5>J}7$LE=ov~OM5l>G)gyxa%l>C^6wdg*9oh+kvbyk}J07iA|i_z7;G7p4Xh!UpUS zV*yQ$b-Xsavwy`wwEz_d-gn#b++Qv{%>VL_`o5mXqz+zkg>BICSy=c_f_Ib6m09Qc z9;!t6hf%(fEQ3q+2>q$>pEqUj?D^)}^t4#UA7ol<7jZ|E^wT?*Zuj2pwA|)AA5eV$ z$_V9TeEiN{U)&GcB{B*TO|1M@bGs?Q=mA{`qLF6Y3bTDV-(e8#F&XZ8KDmMku8@in zAB+)iWP~z9^)q>5cEwPB1D+o<0G|zaXuY4~p^UfeyzB~asMMSd_uZ#AsYJjgAeBVG zJJ*qqxwph#Ny`Xm7jsw?b^{|#in(bGw}X#>2!KT=yVCy_tq*eTRQw4W+rS+f{Yutj8c zn1fQrSvU=|2b`@R4}riO+N1E{!#uX#CSn=1%ARZ&^e@*aV}QkaQ~SCOxP}>fh-r3W z@GmtC{J&m?A%vH%M{JC}>PIu*2&P8w@^ymLTc&9xd^BTuG`8HV>3U?>=l3#A2YWcU zxw#ieBr=t@lXR9nd*#X0^T3-*wem;RR4-okV6Dk;NZ^W_5gkbOk18xj<6l(s;M2$Vy{r6MPLfLKM#<0$aiZ57MoXfBz@(vxFb z-g<52k^!6`dH}zX?Z2W;<+)EDT zTi>LgN$}i*x~{uBocQflbT&hGT{H3v-%Gp#4RIkQo3;MA0gxNg(i6)&g(j?He;zRt zEzNE1%4m6elDi2Am9IHhbfv_g$W7Q#Pv76?`B{bxabyx#T9;fw9j6lcgMhr5b+kfgvY2Vw`6h7(z+ps1kS&z zU~YunKD~(Vm06pp@QX!aSO|uY66QepHGnHNU9sgOBMt|N?NBNfTt7%s7qChs!|zU( z&Y;BgpA>AEZwf_GdQZbpV7`VE$jT)YF3eSl8Xm-R6(D6XYv%9nIQN9I)NC~BDMteF z6C&_mptIrkRFU3ZN_7&M4gTI|rE(e^|9KV)j^#mZF9#6m=aG>4<@>}JTXsjn)ifgA zBX z++_Ths4<+Kxe90Hb3L-rOne3tcW_gv4&gAs=1wLoaz^Tm^v_EC92WxZ7>!rS&}L5RY3 zN~ld`molwX{EfC-D&QP8-x#E7-rnxBACk}~qvh$5PQlqNf5do0hicl&ixWZ$QX%O!*Y9*{EPg%3gx91iJWP@;GE0S9u$g=v?%6fz+!z*O6_sC zqo+TWqY?D|gGe@#gxh>Ibj5(1>6o1ozsVp#N@f37}bUpGM1DkFt+aHk5*hkjPhgP zW|R%;+DDKd!3qPBufp8}?RsP2uI;3RHKXH2J$yHFJGw{CtEl_rBU|H-_$XQCOEH&| z-ep{aPP*3Rt`%FnW!nL0<4YhjGwF=_PkCKFS32MMShA{xE&kO0GJl@j#FpeqdApcLd?5x5c$PWKeW(=KLpPj>(8 zF*xgx8qXK~b)FH6!%gH1U18~G7NnV0gborP>_2If3V}HG}Xt^t-@z! z&GOCDHFG7k6E0_u78v3@Ew+ub-p**f7lXNM286DN5iBY)(&dvC(*~n6KBJD}aJJHO zr)Sygb`V| zh7o+Ko?uy$u=GF)bvxK2oY3w?s@J~p(~E&sIQ&oCG7o`MUlT=UmUc7;0l3ACNRC}q zq|tb5TxaK^DKCiE!==EDT)aI~{~nWMG22WPk>Bg;<{>6oV1$Wi)_w$cE4T@0@($S^ zbYBTRTHt#63h-FsnL6w1-*Fux->S+F2Pc+FFo4Fhu6f{CcEji{)AiWVJCkSgNvKsesgJcWqkm7itVJG~GIBD0r1mWJFa zA!M276_{aPlkt(l-ze|dJNa`POr~e}rfUj8sRWiJ54LJvD74y+C8f<&yG`aF6~BeA zeN?L=V;l-L{Y0f&XAWY0iB~)ioJ%UEpP`;6DVaqlH&)9)MIri1+s!U1`-WCLBMmic zMiWwjMnyQ-;^W=((|IZ{H5w_S&iG*m^@q!-qT{9J!{jqZOe3WptxpSK9fuTlZK=x~ zD2ZqLi#-K}2tiR&43J5sYKGTyOFvGf2CEL(lJ6Y`6mhDmm#eOr`OF>hb%>o>eSVr% z6daMQA99e2DHEo*!e4aPu58AKM441FAz>fznX<3nc)7^rUOn>;ye9?T) zo_IWB-y-!kkkaQKF3lITLoet;^;{&F-;)RmQ)h9<*=(FPMEBx1NECdRqeDue?&gB)9T6~uxq6(S^Q#`3YCTsv#~dm}sx zwU@!E+YJdZ3PmE|k5wxjIT-D-E&L6Rk8fx{2*37xr|57_7Ztmq0;>`III@zq2*UHW z(IowJDubeXxp!-svl;Sf#l5hSwG3h&H_=CbMe8ajf`r@~Lx?u0*i^KH!`}C8)60$w zMT*f;E9@L!-Ex`L+<+Cue{I?F{mBuh)~f>GIUMlm?sHR^`qaapa7&$&dV4n(@G`fE z_bJ!(KH(Rvnh)S^23$HX@jeF{32YlUsJ`LjTvNW78z=d8iz6nA3@^ETzWmeCp9`4* z^flZ$+?YF*w8f;`8UPTi>BKMho#9e^Vz5ciNh!4j{k+BS3ilAt{fr*L@Ikx9a()Yr z$uXx4o7jvJd^<9ADx(z=79Qt<@_&^(kiZq@I7h`k4<=)`&Ks%pQBczjA1o9W0%Nbc zdkp|7w8yhhljX^cwY&Lb3>98|h#wep@lJ=eh9)SE*_p}698{IM_S1lC9)DaDIgqiP zRwL~UnGPVd=5B^An}rVFke_NzWIUFEGZ(?$XlN0F{^YglbDCpHA zOAx?23&Zju;9F_qZyYfQL;4U7Wn_EbF;V>I2FMhyL}Di3McRHEBgv6u56EKlvX*~kfr*@n9j|89Y>mrrHfF5)N= zL==^&cfyoC_j|EvC0~a{i>W`PJYoNn1zFmUqM&c^+^`cqyg|mX)%bs~A-J$4n;g#< zCluxUiAO-|c;k#mw%fn*hCiz8d@JmMb(y{XDfI1*mE8q99s6QSdv}a&iIs6nxpcGR zO!*|Al2xOr!)fy`~|PM^N$^Yrpgsq;e|p7o%_`kvJ7R7Z?Ao zfHb7ryo{>)3K7Z#M9TAv@;cS9fy9}(W|2o(?xfIew1i9T>x*o?v=b@xFTfkV0$7-N zj6%ne8E3l4`P?SusbQxiCkm0AJ2D3xT6iOzkoHB zZy8Q&0du)8@iWc?@gw>M9Uszsel=|ziBY1=KX1Np=Nq_0xjCkyV@9sp$i?STsH4FC z|M0uN^~fKme?XkzzhXwJ6~ulIj$o&7VLB4A^5KvrRa8NcZ^cMR-pHog3)kdI@ghB$ znEVt{2S@_t@m{}q3NA?$;)rCkBq-+u$!yl`j8~2sDDl7;W0^qqa=!srWS(rwp@D2D zS0BZ1Us*H$0Xp?4y&s+e<09CSW#o>UkV8&;74W$O)n#Z0uG1+~o9AbAdCQmoNk zTfbi#0vZn1VMl`Pt-VU$`W~=&MaMaa;PLHStDM^w&BXrti85x|QdD*{f@s}`zK(;h zQBb;ErBggjpIhTKfLgv~@5J09Fw^SsZmsS|em20J6ob8Ud*t%{l6Kzl7TzwmYB1v} z)*$5wv)Q%Qoh_i1)w5#AecCQ+K-mo1ts4y^_tDMj^vHI^l8H)wi%j?Z_-&JQOCrX; zZ}2*u^_$=(ssdF%9db@T>6Fg`niwVcyL9I+9lbAvqU|%z7jAb}kuxVTbZi2xqlJ;e z_lzg7jP!H^GVY-JB%9+4Z^WO$(L-nN-}6b1Pk2t(w6m(=CB(^Z=#+?(^~d#?EHM5{o34Z_zH0tw$FE0-!!cjY*=!s{#196HM!v<^yPD~XB+OEfz)(;cFGwI+ z{aOUQT6Xmms9vCA{^W#(uR{PDsT^7hPjsN_lSf`SWvN8^S^TpDBt`%JPuLOzf0`Sb zTM-9nAh3e*oYL>Wt2#gjw2V?08MqOEg~yD9FpM-c>NiQU$MAgU6is(0LkGKM$t}>t zO)9h<(AF|xjs@i2EnyamjhrF0ZI>ckT{}kOLkoq^jX&Mq4HIWSTDrU3bR4g|_HG-; zy`E<<%1ON4-hs7#k?luS`u|ylNNu_TsX#s%=iPH7 z4YR^1LZ%BcocJF12bFG1C%mZ~Gp^?|Bj-eAIL=IdJNx1)z;HZk zLCh7U64F9QJd1DGskek#yqI}Y5ksM`&gb}qa39rdi?USe1^b}e{Y0TEH?`R^e0nhd zL8+O{(c$adI536{eI)&5V5%y;ep4@vnf=0EjBPoc&o$y%v#buO6U5ayYK32t~%s68>O*eD)W=^B4l!7wP}#%o1tbX6CiOUI1AqPR14P z*xWtDL#Uq1cp4;j(CEFC6hj;CAt)eh^W#r?s?dLsVf6~eN%(Y19+72IZ2FIBEPncL zZ=5sMn=NAnJ2&f$gESQG!26yvUFBaO;pa*ZWTBOyqT7|86Wn$b!0ka3s3cRcA}y zECB!XMk$p43$DgtTUd{0j~Gwyb`Q&l2MFyPiB_J@(l|5Fy8P`Fc>0`XkzEvr)WOK@ zscYou=;-zC+&)Oi#(!3>y^ClMFZhXu2gB-jbRqwNez&Qz3Y<-+u>6VfOO$TVhnzEj z9~(*Wbs7sWHMWc>1LZ1EoiS4bHgH`rlbb~CaQ-ExS9`u4mXW+&2&7Z4^Xl%jcg)=~ zwXLT0J-IQ+qeL0b;}Dh>0*`G|_z@Mi{2`5X0%ooW<)}VADdUYT zwlskwGn@!!1!QE2gZgxdc8ersG=+Wj|IjSJ+t~j=VFaJDK8dbO{CvJ@SXwuLY#OG~saknm){@NEzA)IdNR2Q~K3&p2_wF}GQ2(yr}TL$<#$gs5k#A!s5Z8a0#? zOXB0`gU%ThJ<_T>gJhvSy_DeA?(keX%vCZ+C z6Ln~;=mcA`+=Vg^UG~RIsb)y{94>ci!Oh_rzL1bjzc+@^vpsJ=M*nI`6Es&3{S<1x-mVs##mT+C*fjEMfYC z2HT5{-^(;oM=o0gTuNJ>^%K063a`^3LgAC_l5w9NC&TP7VCj!(QT>yxgyo!^|+B%*GGb9$=zP~yEtB1z_hDU02ddR z;ep3CHbX<>T5kZDHi1WDv&%WNDVE%>N1q`*{_Gued14erW#PbDc(+=2q$z=*_*M9G zb1;&>hrA_GmhW-B9JNV!QtmCe-| zJM_a!AdUWAX(v{cNN{K!(Y}+APzZfd1b$j+7;CIATDvDRnZtCCMCN28c2~kPn0=y^IQxp6bpS z-itt|qMr-UnKT0pbDTi_ISK5R|M?@Oto8iyy0y%&5WKr&qh>6w;F^!vXA#g}v?}j9 z&7+d22439&1CKm7)`o*Rq8tBds0U{F1y7xu1q1}pa8GUK=0TGW8V8V<%*iC%-U$c` zUJ49UYi?i_(E^$n+CSaFZw`km^$H{i z=ECsEnCs%yx4D0)W5hFi28S}ghH)5uxhLv-Tb`n$~H*xtpcz+S{|xMo`Fzs}+$xoQ-3 zP!F*;OWp0+Q`7Jq6w2*f9;1%00pf{cb<5g5&W@2Fp9&r-<827@xnfNa3NWCuh4>TPzu{l z_pQ=9=QHm4V(mc!AAJ$Jp-i1phjIe))C?v1DX)Q3D(Y#{2w@t2huG#UsKH=ks_0YG z+ZGQKMWLMqM|pD4Acfdg$XCGp3xp6UYY@S&2^&;*pemkFymeTHBR?D?m8j#S^#+gbpmAo z-83jyYJ9`jGr{+JOjNwyH%F4=4r0P#s%WtWeK0`ZqamOC4fxj(o|k4*x`eWThDt?Y zoWi3P$s8ygaa^sbSfuOcJT_HaZb&ecYG#LwO6oz}Tp7qnsMg6ad=VR+gOsE1^yGZ7 z^G(>+JgYLYiG?dRw$p)$Xr5Ofu~sc=#)+~yG+I@(QZnxQKPJ#j!EPTR^^c@MFPbn4#qfnHV_q8!yO&hC89&#d_6e~*R?L%i)Kb#pI13BUsU(wFINL@LB0Xhm2+VtP4j?H1P$Wq7pK57%8v0%W;B(6-Y| zP6SK5CZJM>T9^sa$$i>sbR_8n!;hR%{d_a`^o+#+ilfmusouxv9=EA?NXO^%=XxZ*Pv*$M#AUpe3}>C{uWpH$cs38BeSjs;Z{FCb+!$Db#4 z@{|k%jWc+!bDT_ojLpOz>+MAr)Q?t2p#9B;7({z;JpMLucJA|_MFiQ-F0iKYdH<7i z^me6Z`uIxXkCY&Nnsq2Q`Q_eLnl1e$P$ju!S>@3q7(^b;!-bn0&y2Xcc>eetH4!nC zd$Bt=Rh)pRNFITd2Lyc8A=q!6u^dM5o7Py6ehqx<9utQJXE7k2M3}&2Qm1AKGrppw zn)oe%)>~iT?Xq>!lx$Lt0#|Cb>3pA`QBUmbz$#l`JV!)Or}m^8Yje~2`=sb8Mw>u! zN;ReE#-xj?AmQ|pq@SoiBPSVBU+$EK)K$jLJu_8M;+b{;+V>(@y^avDt6j1lv<|KV zPYAp=lU3-s?Rc-G!6id^A~(KTOP5UFaSsA|iPzIawMUjFV`VzN3-)h%R>c4NE&!IM z!MeZ~DW4x2|D5x+ezE;V6Wt zrdkbCrvYCQ`L15W3WclCX$=_Ea{23Z>_#&@B8MAVsR7mKv*FKZd zVx)Bl180qt?Wbmr=#-za);MhXyS{+L-yrg2pgFOuqz1>}(B1VN;&7}O3>1YwOsAI~ z7=Z0wcXqjL^+%=`X8+RnzAXCFWt0CLUrnqVy;7b<@{UOUvVc_VzY*OXDugdp(MTm3 zeZUC-i?}QKnL&wLsiIy{Ia|pMD+hs)q}69xTU+M7Do}PLKv=eKTYitGVCkxu1@wug z2jR!ArcQumhs>0GLBg6`mYPDnvkP#)xP?W%jBdTI1DA8qY3f16i;BdCY1G=`A~&le!w`u}ZJTP0EUX;l;FT zpH13=e{SXGF$mr|JDH69IjH(^U%4*8q{r7N(~Pu&6@&|x@_JQjPs7BD*f}zQehz8@Ss#4+7NsD&v&5Ns#7>} zS-4GI3%9e=3Ql_UJ-_`Y%2@rY?@g0S+vj1M;_1TYo=?Qg zVQlh5x5y&0*p34DL3svK0dpeC62z| zE~bvS#y%J?6bQ&23#DWP&V7S@^mTV4v~Z%Ri3Tr1V3?9^_BrlG%45fL-7dmaBdY_|6~%JOR{k_bN2qStS5A%mn? zCR`D^&`a)f~X*u8Nt&!lDJEpg3tvz)Vj1ph%4`(FjU^vXr;` ze>wxuVh;j85ilZnN*#6!4gx1_mu35F#6PSwRXjN79y8DWv{3#@ZCVmRRo z%vEu^vJ5`ilvZk3szUzQ70b~JLfHm?^uSlU!dpy-e%b!Jv%Qi}1J`4B(}LgFgx=(| z)Y`uzCtJ^IfH#CvZrDOVA)(TA*FOk3+`oqex!n_e3wA_-pCCG#F4pO2Ss(rU&_q53 zfae1jq|4gDRZx4zM-(1Ue~5;mo%bZEf`|n2rP=j42qMc!tFFK|yZvt?pW%9K2o9m4 zo~-xNAm2~>4p<`lWIStm9j=Ic2Jql}NKngbGcU^z{s1T;=V7#jm$DOg2Si?v7k z6VozrJ6grM50xI02RV$>IwqU3Gm~vDZT@G}(n;-*)h+_+$nQR{Tqq9FtFmmdD0YE+ zY3_V|JSo0+LQE*C-*OzF5e~Ba+4X{|X-z~y(DPlC%t3;~NjRPM@-s;7RIjBKRRyy` znB^A1X_ZKEtXPca)L&Yu-=rb54C z%j@aISD?%myjRE(-_u8%-tKHuf=AQe?xz`EZ0-lI1~^y^u5m65;XPGA`!ari;n2f< zgG$8PD6{f6&jQsV)S0uZ&5&YdYCas#PdJMgwi$lVeN>rDEksdHlDG7}ReIkRn&Ez` z#_q*DK{0pGf|ujbtMreMo0#j8)%}##puC+tIsW{((RuW$I2AlDq80L|lN7HIHdr!q ze(CWcACM~JN~jW!K#x1dGL`@ars8LluNG$Um-bWG)3?RiDB>uY=#g@gipR1FgZ^__ z;+==5STu076Fg@patyZ;go)Rbg;=7xGZ5Q;&*8XmE|#QbJ7gE~=5liiRV0iGwP6${ zpUZjwPT6j+zvpDKcJD}FLy#(1zaa+8%^(vv96LsGnU*S+khl{5c%PrcZ(gHIh?}+d z5vUyp^k55gw1t(mV3h$*bp__eCC>HDS#mfsXA7Q_Gn^>wlvI>=Gn&5!zgs~9YIQRR zIhWhBN-R2lod55xaGG5FUZVJuV=@cqnNv(i>h}yDATs7H0} za9C>hViNx_8u&Ei7Oq1N`ruB3t=f4q1Twit90FJC9F?tP*|Ld2%Y@|ZoqbFGcxXHK zdIr<}H1|n}16cIrWjeDW@4<0n-5xwG@M6~> zdAJ8L!3GMIC{4|QEfHR5mb>mQX<2cXNoMg=)a#=$%n#U;&3Uk^4j(iN z9#B)s{c<0DrdS>Ch~p_Uv#MM8{!J87SC|qe+8O~zh4C$1p$NKUM)Hv~h~Zv@kmC#Y z-9bB~%x6CzVGyq!KoU?KL%eqgtk8Q=8ew6FRCnLm_mK&d$?ri_STvIkH(g&`r7?-0 z{X1k4`&bsYn-$@h{>4Kn^;s(zLXjWgTRN^3Mvw89qRA161kC)lCfHR@)yEuAK)rej z>Z-B1+Ur+37-f3k0r`rtjtz9+kOBJj#`B*r>J!?x|z*VynxEgz|37deFIgvAx7QK!gZl)9Fg+$J*c>VywF zUBkiMl25aIfw?bTWCoBel5~U{+aWyYWZPp*7tiW%8o@sWhu~$ z*nHLx3XdvQ!|?el`)uZL#0bYnq7eA*zA(2Wq6z@LO4D94to$D7Mw%t2TMgCOt@;nbPQIbo3+ zlSE@uAy2`j{N2>zFdmI?-p%3<54WNU?ZOT&>i@?Dkd_KFrnv0KCezl0c2N6bmb+M^ zEW5tW^BLKwkD-7+j)LN@a^h7GYKkh_P3}c zZsUOYX5Di@W~yrwGY`;qf?qyeI*1G@-?$DxShn_&qRNUeqz?9Bd2{Bg|W1Ee=dHUhC zBy%6-($c0VZ3x=*lBPOb^LXtYk8`LHU;>eK&lbR0(ssyn=^n&e1bI=Y4M28u4p@#A zj)6~$Z+*9#yBxUsm) zIC*Ob)T#dEEs7sY9!v2JSKL~-l)WQ`;~zo22RgFi>x=8TJ=Dq1ujvw%R^Lu4txS2? zJF-w95`^^YAMRI&K-6lWd}DzuTJ6z6E1=f0%{?O-Wsi+K6Mt5KSZ9S1X?$_T%YCuc zpT8_BNc<-3x@NTvhUN(kmg5b*T#S~net{g+-Bd|sCCuzJ=t&GGE(a8P zp+!y3_wWDpNVRb$acF*e=R+^YrtEBCFthjray0WJwRE;lS0hPSS zerAWH<^(+S()hzrG)H-rcHN{*L2LstXvfUjh0(+K)WRlQGsm0;g0`3wPwsa_gM-< zQRDbnqs;TU!Jn}f&(M;TrP))AbwXBeyuBNaH+s9f&hAg0P7Nw>?IQ?CVTLXcu3)fkDYe6gU#InHLW)z$oOqx@nYZ0~yZ{q(c_ zG>l1Pi}hv1?&4Y3Rl`5nqviKqpYeIOX?vmGA2z>R;~8Lk{rK!Y)<~ifIwQJL^C-T=ztcs9l1fj>7dYO)Zcl^HJ%C>IyW~}>`)#8 ze}gyC=!S%=q+&_sJL~}rj$_Bh;EXc+gkET=rEOU z4z+)1&dv+v| zX!)%3%=pPBs^yA4k%CWygJo-=wphwgsEJwo>ANDfxDhkD1~KlUK(RAR3m3+%RIr^z z-T8cTSr`QkZG)6bF@uwON=&BeR}2ZyzMQf){4VOP(@6@exKZy{Q}bz_U?~7I6Kk1D zHt;WQCqDBhmpwv_7Uwtpy)YDve93+~Fg-vc7hW;&3trUB}i?Y<9d)|-r>>wRYjzXA;5RR;k ze$S8`2Mw9I)pMwrP+Dc|D-dazQJlLeq3sjP2qI(pQmw>tmrs@MLeST7?yDSF+F*yT z%BxMrW)s;BMcug%56*`%0TNctIfNFYc+B95zMM&^=&7;MW(@m_!-PEt3=pj%_j8=Q zCAB5~+5g3(s#GxN>YP)*V%cQO_SWRVxEmYepM1aAgHVcM7m|BHJ-yY+yks#1}hzkS3QI4Nrw!e`IXTUE^C zXn*cAFUYtid%`}~=N`>$f%Z<;>|X$QJdDyMd^V=9|E0*F$1j zm5{NAlVvF*QbR8%!+jz`?Ry*iQ1aKLsH|STgs8LqQ@5|bs=wxFO#qh4p+!xuozsAE zHDfF#2J5#q{Z;<4G!t%xwxGodWsfcajsu7HzOn5xd0lt^TkTiG-?{Yi^JxXGKRR)|bs&Vk z`?0}n0CG^)=u&Pr%|;|aH%Ut%nxaI-E}Kgblqv%oEMa&VQ|B@fI|#EkTIP^oGgE_K zsl1-uDCas>!D4oQVNEi7)W!5qxiPOReqr?ahLZ9yRb_0SW=jYsC%Y_V1_>xTQ1-XH zrSOoW)sfCYR(A59W=woBy`Hv`2J4;vV$QaR;5Fx!mbSV-$KVp$bLE=Ji3Y@X24^e8phA&X2rxiuU?nj*o9#+f!%1w(`#Ew&#wj1>y7TMRlZ^=ltHtmH zb${+lud4_C0t6IoK?dwa%@T3PJA!^_^&CPXtO(@bP`pX%yN9_c7IK~e2|MK3!X+4- zBK>Chht1<%tsbV7f=mY?IR46_=6IL>m(dBcMP(Mzy7y`H&(EVB=_^M z2-6jzD*fxeRH^suiPx7iwGr51#*xvl?{Ark>_Y^kZ8R}q7xu2y>T5M>y=N2W?A-_g zuOyhJ%AReDi|83nsMSf>8T?T98{<#5w@1M=IGG;v&nj}Jnzz2^&ONAwT$Ul)QwBYM~yND-NTl>Q4{%*~r-i$=!+D0g3x=?b7IP}D$ z6@TAs7whj7`DP7Os!kL^V%w7opYF$?f#REQ!))H@lcxMk)LFeCV+9iuGQsw;_qAle zp>6P#k(Q?yC5YKl^7Rk;w$c|{!O9Q9Jx9tDhkK2moXeIDO5>U1@H7aNcuhLrePg)1 zCLe02?a$}J`^@t#1L?$7rXWl-!GcD;M2!*idPIiX(eCJ zBMGMlQCgk4B~j-tntcs6sv~$5UnN!pl1rUd?&BQ){QCCI(pJCCvG_KgwJ9}nUB>(M zUbT@cmjIpxug4DpT9rssqA6-j&~2Ti>#(L^H{-IJX{{*@a~FOHp~4g>Oi!I@LZj|A zB_dEkUtriM?`!b+#xrslFcRf{{g9j|PD96~Vjm|YP8V^$!o7OG$tKij+LZhb-%S3m z!6e|4wBER_UX?;H^@4@yPx!i*A!*B|5EhU+zsj$X)T8(HE7t5!deb!EJIa05_iY>( zwT{H(=mb(30vyA#_(YL69-Q=pCvuw6r~5|AC@AP?1#|58je7DTIkQfj#8Gzb&5J&D zcv>QL4q4sr+|;LXxmUZ zD66>i7z^yoZHzcj^>*LZh(kgm#%V5XwqM`QVO(tM_S!iV57IH}hLc|_E*dM3AWu~3U_M%xIb1Ybn~QS@lT(~Z z%sFbh6faNVF{#u({YNG|fd$PAm zPieki zon>~GCAb~G$LV74yPHfOK7r9U;pU%weslyqT1_;*RXepS?rXSv3vD5Q)mwd2whmZy z9_!1;WJ$_sR&;$#Q@GTmcMC0fQ2Qgk#SilBS}!l{Cr7M?R|tmSNtA*9t~6)1d)p?U zs;&+ytiL7qa{&rihz5^Ci=!IK^{_yd&9iqKAmMssVW2nm@M;Uz`OjwMO`b>VZ?6US zbJN1=ynaYOBD#Bi?GrAcJm;`9kzIHWSE>}LV%;aF ze@TPEALhg?L`d&BtWn(XVg-#goQv|AwK%Snf^x8+{igwtr}*(C60~QO-Qk?$hzGIy z`T96?OY&&}D@tu!*xvgAzm3(Fo-~9T&DSQKf&94BC^Ap?5nA~pwL^SV?Jqgocq_E$~DJUR>4J`-}pc=Z*sO7ZY(IU})*}-fP|9^U9Vjs9q ze9;ZNpz5$aCQRVuD5}5Ko z%m*4_=CR+<+@m6+i^gUN;~<^__GLDF_n?mfA6RQO1XSeM9?W=Ka^m&0!dElWaF{bV z8OEndI$9*a=YDDNb#AQ~SvI!Q@%R&|c-Z;nHzImY8MN61+rYcOt*&Q1T~1#8avO24 zFoc5-8lZ>vA+rjXM`gxxKcD4bg3b!NI$wmR{s0Dy>IOX2cR1YL8#W6u$a@3o=zM>v zQe4bEKUZj0JxYFhr~?)jw;+}ZaGR6-?a_yC#tcRt8}Jm6fOgzcs>lhk3HT@J<>?mf$tF_ai*#qSIe)U*5){tuw)G_F^WA+<9u=J)O`rEYs{6~rW`!kcF zZa(}dGqzL|xdixeU#tReZ9=EVQdltRU^b82tN6K3YUo+lVXh=U>d)J;rzPx{QJ(@X z@G4@|5`uL)+MHyp&d9%WUt<2P+3Ke;e%#v-f82h;0FU_G?wyx)!kdv zrvty9x~Ln?OcZ{x_rbK~fqm!QE`mtQK3RG`h&d$cu^Zovi)=e8a4_+bPgaxYb1TL!ze> ze)%^@h!1v^^uS99Ig-f^*oa?;(~B4_W~jOMEN!(K`J%Jd5dWw{AOl(RLUnfYpaGT; z1Llo7TX_%df@&p=ByuRM;Rz6_)j%(qt@#=G$x^epT$*x#gDzW|1)T*mPG?L7|7QA_ znJNK9YL?JTl4Bmat3h-~domK-zpGvuQKR7JESX8yfz;Osjd#P2ih$_?nfJI4u zg@-_S5jJJu4oM76{MWs}A4Fp>>1W!$?16GOe{SuwOwg3YQF@S7f;JT&iR)hI5`sgC zTrq`YD-9rum0t3_fkg>typIO2)C;8QD|`~D*8h;*mEurb)WA2sH4q~?A=I=oBTHo6tTo~oANbRGxW}Co9qAk>;Isn$ zcV~}20l)qP@8B1Hs(k(S`3aHygfr}_lZAPG$hrRW$j-Zz*FQgs3H{b*_-icPN|c5 zt&XL>|JjE1+!kdYzDiRmoocz$Y}iBmS%|QBEb9+i+DUa#;chzMC~HF|h$K)scd}iB z8D7luy!zTjRuR7#mZexuRfmy~-X<>n^Tkotpr4(tTUaK_!a$OD6JsA5A2!bQcDE)x zZp1Pi%jDDU7saP^|K2OXm2pXrH_QFRc_K&IIK7zTjNK=PCo+WzizyVCPd59Zv)$)X zKPLj{g=03Kj@46Oq`l2-r^bZH!3>Jps~Ass>XE^SG3SOOnc24RMFjwAQ9>H@ES*I0 zc(b2RAx=Sj*zlvueG_P6Y!7HrjE}@h&WS=1Tx}Z0NqDT`T(dTK92;0X>v{96k3gL}B@lq4A#&W41hvVG9@^{$__HMKV2zPNGuN{1;s17B^FX?V_=_%Q%oG z01M~9CKSj&H3~9<)tT{q+qJ*9XS-p&pngqXo6#U@9N9kgDPwf-d_yx(jbrl0KAkla zY}z>;DNRYC2X1jXHA!;%(Lh)NKiQ3tvdWMSV-GXmkfwI{VFfv89on%KK2OA4n zr1S-j+EDJ^oksTZ_4!{YUdY1ezd%?wBZv&BA6Ya5@;j0*Mm^o!U3YTBzv%l^qJ)5xXi@AnDBXV1)3HXGIcR|h zYCLV$$XXx7kI5_2lY`VHOKY@8zQ==iw(L8lWOkU;jMt6fe-GpEjn#DR!=^)n?17F+dbyRQ>B!=Q_^p_FSye zMjo3V5EJAzZcqgg(j|mpfObx3`?w80mw)suV5>JMC2GcpR;2~Uh@=MRrpJjQ7Ep9{ z^R_4Np9J@b*i$2`Gs6sy&Q zY4XOS(CBZ?*#GPw9IIwPkqhYqTw<$zF799V2G7abgHjKFa5*8)&{K84US|CfRugFDj_?P4$ z(Jx@L=MJ4M6c!2oA`GX9bs`Y_lE!uR@2YF@MVo zZ+In^KJ@0$azJFi|!(g5h9V_;=A-E9!jAR0>xNL-7y(ONuS%OuUl_MJ=@X;PQ zS7Jq;?+D$-09U&l|yl)BnS#n%7Mfdohuy9oI;H*{Wr4ptQ;>xGO(yy_n)=jdR_0Nn^-l3& z#gC>vRnw&m)jqm;FayjtIQ+l|?t2>Q!j6Wt#WVX=0BG87pK^kv9Dj#hL*rd)kIxSAXZZeU2nE(XFO^)qc_K0}=P>fyTifyAJckUQM}66_VAHqfo5TejxBr9Pg@CD4UGU|KyH5o7B z3VvSO2$2dnn3npxWF`4#(mNq6_YT$MSH^4YE&TCKV%h^C4PT5y58}ZmlKB^ea>U}>;;R)cl{~09`=)+fCHYy|LqPW_#{+B~V-1ub8y7GH9ec_hGTi@! zyqzpu3+|)nE`tlkyzSYed_mYu&?|ISHMH5|K{Z{!DRmlI>#a2t4a;P9UvELyo;sXGq& zpBKbb4`AIsa;UrD`0_iKY$rDk_P@QK?SgoZ4$5gX_$wTQa8t(f%z?ab2|b#*Q-Nn( z%rUza`H~*5?{7}|!Ax%qn6k){BGA&aW9Yal9k})`=LVy2ciA{U|#P{z)k+DD`D)5cqz;_zUow>~NU9&eSQDqUXGwCf@eC{u{ zG9-59VtvdZ%aQMzhVt02ptR7{w&kpX@j{h^5B*=S#{vVu%mrdFt@W~7*zeL zuS_-@kcs!RGru#vI=r7ejY!n(c(5^Yq-U%+WKh7iW!j|LVj&qIyO_;yJ?gpoeWcE^ zb<||_JRX#K+0=_cu>@ynoIh+Z-e-KXzpraSfGM<28)0|H#3pr-9|to96H9r znL_p&|MW*K_GUZOT)pYXcoeA1|JC%On$C1vT2-myj5$@M54(D@2-7d0XPr*J=i-GW zG3u6G)(H|2)$-MjpI6<^dzI1-r|m3M!_Ir}W)Scz&U)8>`QxVkus6+`k9=Ir@nvi| zy(mxwB@rF>$m?H-2{!-|d2{N_J!04zx^xhMYDY6(UT~MY|9c-z+6Y0$S#}5Ml~XJL zR^gDgEscmL8)%D>|NZ7>;BZdMMrc{z~tSz@uR_L3J?m* zIH>Q~kZ6TALvOPuE+a$tw&M#vu0Y@Z`}Rwj6=90=>FUf1nI^)*Sbkl#cwRwgFLo;J znNBts+@w`kKgjERAq(dWJ!v?v1pNe+AstkLW_k%Lf6k+{l9(hi0&ouf1Hut%y#H~JaJ@> zV({N;C+nk~(g)%F>cu8!`0Sqg_agYC4j#_~t&L4`i+aLLaZ_p?TTLl#Vdl-xZ zP86+vlo=?~JC%|-;N(0mhgi2w51bo=Gl*t93imUk(~Oyq2S@ZOp~q6i%^CqQuE`Zk z(t;wLh8=_&%kI?vUj|>a461-g`|fTYbY5x`+K0xOy2T3%;vzm-RBwjEq7nc)CX@*C z-Bdg|Z}t7rA@(!{ERaS%J0tYZI07yVQs=YErU7aE{BnU8zWxQIX&O5;+K)?AvB+7W z0@~<>9}aAZrJH=eVHxy;LrTWGftpMQKrA2REbuGE zqlUZeHfKfbF65`oGCoK8Zin_Hs7~M`aH3pN+~{$A4m7E+289e2PZn{6jiD4~lN%xk zqj6Ap+PG3i7F@E(#CV@P+Y@2JJ}#sQB5Rk0oNBi?*!N-Mu!layW8<~C(1bo{1Y~^~@7pqf zf+N6Mph3Phf_5p{m&i=U#H9A*U0vCspVb#Kc5Pe*!qp4={3o{lunbdbYfZZ~_Y?wt zXK|CjFI}F6=mg~l?W{uc4Sk)8l%H&xFC5LqTXFT-f&qxcm0lZ<^Tf)ae^yScm%jsq zUp!;>Xn-K!i8{K-^O;n0Bxd@*n%edpSZ*0#{r>jNm4HzH$vb_zEA}AGxuORK+UK>C zgNJb`z}!jGTUuO+jm#X&=y? zyufFOcfRBst|Wi!c2Ch_5R&1^ijVFiHd*@B=_esuzGX5K1!@6Xr?Y;B^r!NHss(9c zHeprSl{w?BGFNBSW%=XYNiDA?TgsXO9`LRfr-XN+if;}8f1{9pTjxxelEvJde?7@j zYV6Ns&CxCeE*&uaP^nFo-VD>3W%!+V+vc$}s*t2JkQz%vvTeTM<`Zzl8t-IuVjBS4 zj(6&SaV;HHkCu4g4x|tbxcrWV6U|JXn>hs|kgnIc=6YTh+5$sr=3pEDkMM)_M!$tM z7aobJTotnCCJvcGj}LcaPN&v!g#1EVX&-J(1Ll?R@>*x*bYS``1buXNA()-h(&cnd zg9GRCWuMBo$`_D1)~ld>5zzb8(ESQ@OwPx#ZzM`aSnPDMB57i3N_>A3Otm2xqqTiC z2XQ$0Q6t!$QX|-CSgIgk^(p^0qH8m}?b1l-DQ*P=6bM-l^%wnleJoyzWBsPnH~dx> zBDtgpAMbfMdw;O*1HnXE#RXcl27L>~Bx+3F9@Dn8U?gS{ovyR{#<)ML1Oj=R)mgAU z`cI3F_4_;Wt{E;AhL0Y*%G{&($eKUtflv2mFS{uCbzXuGVG8Qu9E5KJjQD13-AYG$ zyZOOQEIX)Rn!&Bm5iLCLd^HoinZH?TJ7q2)ZXbU0O&96x$Gn5}SHTCgoHy90V5M?T*CvM=-U;L+-n6!&ZJd^!()w4+kE z!wu?2C;0w^-Rzqmf#H{>!?sU+^l}~6io@ePLXKEn5uMHctgBszpWVJXcrN+TVMJ0w ze%Ubz%ljeySM6u)n)2yoF0{mb=N$qlbVWQq{wP(!<57$&c)9+WPl8)G3ojgFP9Fo1 zuj#Mx^O|8vZ6X8#9V1E>E^!F^Flg-x`EZ)oVEaAaZ`V%6KHO2Wankf_zO)?+-kWl{ zt@t3_U)g*It5M9BH19pmawauz(~HEp3Ty?cVQ6eq>O^u^%+niAJ3h zT~hGg5nvkUl=sQ+cUKc`qSMa(zmCm(@55p|+{LtXB4-Dsm&=H4GP;jWcircGRJC7) zl~6V3dLtUI+5et23X5&;Q5lHkd^L%aoV5wK_o$kp&y%~cZN-rn?StRGU+P-&dGR0} zGkO>-zd0SnIVIo3=brw)q%^Ejjc8voc&IRJQY*bE$06k^rwhI9=i>k`aXl2~@;-fV z_UjX#X`TRq9@aPe-xSQ8?%IoQ?&@0dk)}lopouW>SbBsdI6usttxfd6c9O)*17Ui2 z;>a7QNUxv-1bA-KiUpfbOm{B!?{TUibRBJDI3WC?r<18?Sb1}fir5NW(v!xF|1eFT zOgFg+L_*UJy3Zlf?C~Vhrf-%-Akj9vo%@I>mM{SW0(#v%* zYvdg!OHr@HXv^t~Cob2u%7$nj)T947rBM;kilL%Gu_DQbm#c+=ms*BmKl^)btjAX- z9ex3_=yumbA*Y=KoZm9*_)Q zn1i+g66c-Mo-YG}e>aO6=ng3lV>bQWYY>}c@B>ViRLx4BfI}31BZe+M-#@Y+QBcjl zd?b};TEm6CAyNn{jjjTYHwBCp9|Dd0(}3I9Xc#qJ&;f!w5S4Rj@9j+p$+>#v{0b8O zC_ux5^6h#T9_g|1*Ilcda&x8P=1~dv{~_uvgW7uEXl>l3xD&Lv7Pmkt#icFouBCXf zKyYZ0;9gvdON+a^^jD;4a4Q5U1WQi-@4RQu*PZ<#voo`EKlfVeT2tdt_I>lGdMT3< zv7@Q$O*sbGHG#~{AF46386EXHLAXGq~6^|+I!pQ+wM+drbmguQZ|LPVS_2u zr0X+3G+~*45x7{`ZZi7l#fWsCO18L-;CU`*Hk<|a>KVgSrPRTH2`9?$OJVPX2{=it zzK=!<#19F5M+TlPVs8kMYi9ARq)WQq)EIB&*dzWmfOBF2p~ix$9{rDR7!iv`U!&ID zA|B}iT;oUaRInt**p$@~yDEFW?}=Xs;3L1p2oQq8e-a}~>2%&b&EYg}n8j}+U)Ru5 z#DrnIdjRr1f8jcL&nC={800E#Spb{C%KiAeF&_)vKm8XsW^$=kt>Va;P9bg`PR)cw z2fdH~`{nj?1S0he27!Sko+c)^Z0n-3+S($@U(~554Yq$b!JIUj&ml-*RgMV@C~~5H z(9}%c8`YF}NJFkvlx*|KoCzTuvcTwws;IwTy@D~oKh;kMTd?3QGkd_pvbsOEiHmrg z(ooFKsX7mPG<5TP>SoqJ7dP>j)8)Tiv)%EW;|NMPBP!IwQ-`=tc8obr^w+4rC@~B= zvRVdfh&qEfewRS1h|;GBJ`;uymi<{OO>&L0sGRX)oV`eVg`W1dXrm#K z!Rb?2#GV;{Zl-|IteCQbXIS8MqGS_AP>D;Z?*}b~{~CEqj}`SPe^#_*Lx4}?-N=F9 zgX+qHOcenF=nZueiJj~3D!a}8^WPyw#@p5f3htdbF)A$QL<&!F?9&@}n*tT-44(GO zOE-qsM4w@4_L|B5K>no-(Dh97xQ_l(y@}f=rGUQYlPOMJ^Qz2VbQh1_{g*yiV%z3& zU3BpuFM4DRJ{K+5v=lP#4yp9hxBJFs1&hm76$jFF<#%L@Z9`-2Q?e)Vc*O%ok9qM{ z4UE$tVUeW8SjHYF|J7kf&oi!GD!IPB{0vnw3;e2vYQ?5!;ieB6$`3P0_YpC&Z4)8=Ximu4 z&*Z6Gk6_V9jwe&aaOuK02b5Pxj5EnU)3byW1&7$8_H{cR); z(B*au3_N>FYpsAuf{9B)twIV>B^bpy@ca2j42=#LGu91G4ADRxn&=n`7If#G%S92T2>RA(qeXojmtYQjuSq(>s z)-o)47z+`g^M+e8-!>a7Ti^tpTY_48Fr@0JB~xj%>O*KTh0uA!uPR`d~h89rK{QT z&ba9C_3gD#Zwsvr96StG%>iZm zsit=Wgoy2q_?VyV@BPE3B-E#VZ`0^||NblpyreZ^dugYK;}_v_&Iivf#KsL70LV8B%;d%rl0t&G6~}h2YJcmI*NdhVh2aNu zs1QQIrr*2qMpkowL_3nC3qM8}&i33)PQ4`Uk`It(rSXeZxNeoc9NulZcAQ{BB~bP%%6L z;xS!6Ex{xj1*puE@DMFk?R71{nWk>Ac1{w1Hn8vNl-4HXFngFO|V0jgz%)?3o(+wMh z;U!F*)8vy%_rBQ^yX-qY2!&s!L^L-r+b=pe3Q~sYhL7bbx`0mI+j51nz|p(AnRt+< z`m6mZI|Y}dzEX#_2^nsayV(3n4FAz-I#mAUlW@&)y$F&?7)_PEm%iWsrqW@7uYrhW zRoletBIHidP>ft!XZSnOT9}8iq$tC zDTSzxyWc`0oE&F&nh5W+2|YwG3!|nRpTLYsM6wUr2M1*oIuY0_0pro-j@YA96trW+ zMez3HgS_dIh!I;V|C8#!(}|LsEE-c*Gj|a7iuVQu+R&q9#s@+97B%vS1q!cv@_jJg{veom{ zB({(R0JD1Fsq~8`xrs4WxPdtQNEyK)54Qsz1)*lJq@7@E{z1Bm%-B)IYtbTCw5o{B3Jvmq;}qvStHnx zRY_kT;~^GSPGthUc0RDoy2JR?U|Y&As59olSYLU0r%JA{yk-1rzrDuhEc;S}E1iK4 zY&TPLp13ujj-cQpOy7TC2sARsLHzutg$3gPRcMeXXy7A1J{_J1n&>x3{F7@TqW5AUTV{67V;DYEPO?NS_b5ILc{xrfIuC4X{h91fjb!?i0}iWc1-hU_u^$ z$J_tDJ7`MNvEcy2IPtiSAaBN$ckaL3sIFS5Baa>+upi!%>hAZmJ$LzA;O`N(UF_p( zcgZf>&LZ39NAZVQAi=cmcZPiUwsuN&xf^uCqdJkgAS`~B_GpL^kwSB;>_2@UOhKGA zaNjd=QVr!Y(;sKSb?<>T3p4(rg;lX$y~OsO#`YEGzu3QQnxF+(|5*Vb=ZV)rSwb=I`VbjjH6#D_+K6>^1Sag z4>W14t)n zF9Zf$Yw;2nv_!cJBWQ@gwP)e*=A*uZSk^w9w;EW`yS~4EE2BcRR!c|vp7N|@ z^fm7~X`aNCh8c_c!oS=woH;vn1$bZt z?rcYrNUS@|jOYrMWMPuD{+B|SCG|N<$omY#OW@bv2tq96YK!C0Nky(SPJ9oa&7fJE zj)l)SYi#ytBK=eZr@vrJPb*uMMH|V{`hQ|$fEg*so2`@zxr!?0E7)M=<9V0vL%1Pz zrR}?{u7;Y(1NT3zB3y-_<{L~Z41+LCNS!Ik{t(3MMoRJ)oH1z-wcKv6eHsRSdcHUS zAmQNV@d@ntOnBAa{ZCa)qmp6l$soVe+%{|ZHib5tcmK z#3IQOlf+`}2Ko@_|3=uNWMkC8p^Al`>zB=qTBh@ZUwnl=klgtxqMOuzGox!(KH7!hQ#0xw9fNvwey-c4Xlp^%|A#^h6L z9Pxz4N4YsO^b_3V8D7k}I?$+7Vqnx0^hg-e7w(5&2PsKvwJ~#{5+RB|F`w>3JqL$y zOSIQ?h82togXrLft6-(-R`ey(3x+Gvi>zwXj_j7VGB z5Ed(q;jzRInEOMaY<%CkcZ{WOU_*PMIU!-V8RTX~)2ww;5~XkoRpvf_TI=_gAp`|` zoUyGZ#XGKm`2cWh#+cwi`IVNPF&pFO6=-9wF~bY0^?{}Sa#qz_GDc^nc4i;+4E%i6 zahiPgqY{`)v3{2h6Ew0yn}^RDCkZ%daP`^JQn~-rrz?_4iqPFAuNo02ST7B8mQlq= zx)vjo8BOfpO=HYLaub(*o&&E_z-519^#Y{EqvpI+=`<1L(ok-)d|)$ATS6K6-0a{%bG|~@-vzF+Skc`MQDHmpQBGr8# z3fKT@sX=POUDP~#*(c2NkW2llmA6EkWbB&tD9=$_%W|Xvwy0SIw3VYH8 zqfrOGeg#GYpb!)}0=b1yXZjD~Y$jYjgR4i}NR`uTm^wRWY9)BwqGqg#NjWT>1XZkja^4 zgHyydxAu=|CS6%1h*N%2HV9(l>P6yr8N(&O?_gBPgXKW{!YzTQ2AvBgZXQr|f?o;f z9jp|7omZHY`Mjeh8-79G7y>K3aWXTGZykA`4dn)5(^rCB288etQCnzz@(pcSp}K*O ze~F%S2TH^rokbyTeCo)wm`sE@?Bxk6L|-};1HMH#>Bwk;gE;N*;X)XNKR(7INE8IZ z%EHmg{3BjCcO#u2tszKD1S6-o{pa6lH&s5konH%&k_tAL`Ul1mp6`J}sj3 zS=3tJ_wz1rm4ui*R*e5BC1C!o>b^&-g24A2HPyq{TS?&;hSvOY*fc5w0?y=J!U~xd z7&f^}0|R`NO|((R2}%7;_?--FdW{R<8ch0-^6+?v>P4@W;V4FoH;4+RT6J=&xM z|M?ffx zY6qidKO#n>ic>SID8y}{lEMC6Vu%PpY;QrgyRrL)ssM#3A3 zm!YO^jdR>=0efrntjZ&);giD06gNoOi*Dlm6*(U?YCP$83v7|kkC>Fcaxt!)qeCHg z8}TaBI0xAWs#CVo0-`ujW%%xKq4m3azW?g8;vr)?u+cr@6fX(A)xI1;Kd3ec%JHF4Y*l;x-Z-{W;YfuwoDSa;vYl0_8+Dw zGJZX7JIY;g6BR@mr`8Jb|cZ|xgH)E8`3X3*unVeGFx>|IEP%EvpiF?;M{B?Aq@nQHF0kzO;3pZ@= zZ66G8+vn+b6)_*)_)Lh6x;G=`)GWSGgjM?7KvuL{k@0wn9J?xnT$A`<%dMfg1j{&H zM8g4%s+EBSz!Qt=ZU3&g+NQrpx(|T961uM;*;XJX?7*kkqzsaUZ|%~)zQu&B+HU^H zhP0<8|78QFN$~Yq;TtC(R(U)JO{<5gcY3cIg#R8nlWw_`90UEEG4ej69OJ&MYz`AE z2Za!dralGH^Yie&Zc@h<%h1XPs{v+n^yx!Ikc3T>bii|Q^Pr>-G*2mn^ry+!(AS*H zqA*@4_;t^D z`y2mIV;V9EIV6)qdq3NRFMIkJ>pq2Jwc~C#^fME2mxW@|T-*!!JfQ-v1aFSmur9Bh zoMwOpNEC=+*Mr)0VTy_I_VINu=eCu8vIo7b9Uj%14-jkad_*_3 zj}>GDm-wppAHws+yi$)>TVg>pS5p7g*vJoVwU8jT>Yo@SN~- zcu}%vH`K_Y1Kf3*&i}#|*K$=E;|@Hm<^X-t`S#_>cVxgR4?o6WlH?P5K0e?`M42cw z%#IqaWQwH#`CPGk5gN)z0S*^x)q_8U5Lw|MFaSL7o=|_!J++$7;SbFNS*0Dx(QaJh z+}NowDjhNy1zZ)%bqsVn;P2PVDg%Nci(Z&kN~v|)7Aa0%sOUC*Dm8L_o^U7?5I!+&yL08kOyD|FP46aV&=D`%SkYr@R@ zul6S1hjCfjM*?oV@b4jwQ9b)##KrPcJ=v0^52|VZ&V7W4{8WRoSX_Bor(+$o6CPNQqRq-L4XNB*VL)`ZSv5;k=kP58qY z>yTw00>p7B+14+c75|Fo_VGojLy7K7btSo0)xkDHt})?6MiGoI?A*M1<2U+a5pX^S zxZHW%;k8lqdjsl&cD%FC;4A?%9cqh?$P8-oSVmCQnHBKx%m$z5Z!c&)E(6gBiwY;a%-0R3?b4l9^xAaf`N?}6 zNrujMukd{CMEqZTIDW@WXt(Y)RSKrJUTBs`f5ElA0tqaf9lAfV%A1v%`lcEWihcEq zVJD!RLRSu7)w@RsrYi+~(*e&R$Qo*Tgop5>`Lu-wo8JtPrfEoJ#w@X#E0*l8-ro~m z7t_rSY|#069oQT2ml;x?Mrh}Siv4g~5{=!ikJA6dTq~=SyxXdyuy4QcRd8TuV{^s9 zOk1qXCB1=ji9x^dEOF!4qjl`73*}d_WjwfE>cE?aR1{!bhRJe!VlEWunsXc?;7J=;zmwhH6G1aED%x)4%gqj8X<_j zOB9=)8oE>nQj(xz>$WOI!CIQ(a|l#2LOV|)yu_ZD`uY0~Nk+Y%Vh`loy5_!4O1w}n zEn!1;e?M5h-cM{9JxR?XFpocdi0?3`2)rdx9}yF zZu?#)Y6=FZ^G_7#@iltW#M1#vf*d3CKQvNi@NV?9{J|TrJ*NG7>nqh=X0X%GZS%b5 zl@9F{M)EPn;8x>`=>plY{H1y(YJBhux5yac4+k7D41X7&g~G1Z2_f@jAfkQO^DT0@ z$iV#rjhrcNCTFnGckZr|z#LHE_K>XV)s<(A`@rjWqQ2+rzq<(?G(%wjCb)hi@9vCW zn-v`uEuJI~k%&JI29=A}YF2S>b>HJQtxUMB=uuZyv)JYZwvyAZ22wHwM#g{p{rz|U zf=iy3Hh;{bOgZ%xOzqXC^aqzhjpouvl+=+Ct(_DbNh+gHEnuQ5zaFXh@XPzS_SIZ} z#KT8Aqi)?|Icm4#_9Gor*!hbjV3LSs>d56>lVOXYd&gKwP~u5*WWiTmUKLjxIUl7i zQR;}bpYjH^;uT7Co=YZ7dxQT1^&IyOBwoZ9$>mP_DJME$I2TS0i5&CtplDm~#GB+WKY6ODb9d9l-q(aj#}v`7uW8dqn9> zz%;8sEL_EcNi@u#QGnMN1N)o}%EgRS$x-6`M?W5V1>AQrfe~LiAZH7KR?eYHcZ4Hf z}Q+6bm5%&YBM1gL^lz3g0n6l6f_ z-dH=NT4rygSIRKnJ5#K*IWCV@#_udF(7T$^Ngd%q6k1sloB2Wivlrua5-Tq?PpNZ; z0rkzn^Gi9x?8#S%09U*ml>nJ-STk$`p7v9iKB?nK12r*BcHja6(Eiq-@9$=q-s?p5 zq1s&JKz9p$Ztyu2HlfgGvOX33zwl=uA|BEwAq6c`g%q@}*|iIL;#x7`sA}sD$9$(d zzydifHLCiZsz&l>#u@d!gO$ZDfUFaYhA%R@&f6($W(G!lHdSA{9jZB|O=?0zhnM#l z3|m}7?Y&jJd{43tLO$B>q*Tg@*uA9WqvA76-RkQO*3o=3x!Qa;#3q-aA;OWgTet=6 zKFg=EHyWD2zw#P4z*qW0DnI;2ugPFhZ>69f;u|{R{ehnRU!OGkDS;4lQ?GzoIB1W% z_QTPIQ+@1ozO-4AI^x<2Y5SJ-FX>9Z0lGWaHwoKczh5vWPL+BDu8Tn2_i@I8&wGr7 z*j1m``Tu(^A@uU~!VLerEo9kBn@3|bQ6dXQdf&+}1K^MGTZMll*!qDXU;HEnql}E( z$;YNhlI=g8m?wHGlDp{#Yf3r}^{J;-H1k{ykt)0ZxdhH`^{wmQ55xj;G?BhM64<@Y zf^UM;VE8pJd@ry^$aG!x6woPQ1YV&5{bpveDACVWrjkTD1Y`!AtwTc7hfZsx!#cEQzpaaVvx zFsiRD1nt2)%?$fVfxXZJt#EqO!<5-WhjaVFYp;Pn6SJ&D12TB%_=8nBcHCn;_n z9M#UyUcc+c-F*;<@ z3qHK78+H)%^h%QXI>0HD6Ldu+*)D{m!#}mPRrxJ6CILw2VceyC1+Z!Pbh~RL)mz|Z z#>G50lcMLXM7Hv`+ox$3p3+|wt#gvDGoLi>M$eEsZ2a8xCU={% zz+CkfOFfpr*>53NQy{anW*DW$T^t7&Rae2$r?qS1!llafYq)RD2 zjUN;uG)d(JOS@yF9CQlv^s4rMC|rJemH249W3ZaM3&zzu*;bZQB6DHi*L^2w$U%*# z+4%R#&gzo%4iBGs7$8yM$(T)JD0Iy5shjl~Wsob;{_-~3K+{M4Lpf=4$*gl< z+PT)yH~LgbQ;KVuC{6_;f+&q{b>S>}2q?ZY+~?$>o(Z!ZbXs_5ZXKdx4Z;2_2F!~T zJtqbHJ4QbKLY6m~Y13o>@`FA59WPr3*Qq@XEyea)n`0QJBOOfD`bjm*f3-*QhWy)V z(Ah;)-pJpr=Z(g`O8pTgch6k`Y85<$P<_ERPlA2ZnzhOn+5%SVyLQ1y-L#VB@L6)n z5DE5g&t5+%m$~za(q}8OXbPhq&G>YF$}}FN?*%VQbbzRfk-k+5R=3oJsjPXb?qh1;9cb@s~_QCTvY&vE};_>(6DR=T&hI4wxGLz=(kzjE!?2d z&u4!oZ#@W+D2!Z%m%eW|iK>71s^DUTd)(mI>(2BXSq~i|6lh-X#Gb*k(4`4x#BW6w z&UAr9v6xmun?usV<`BPnzy1Q6A?+w!=;LANNT*$(F(%K&3W6q=u#g+Xe`QARf@}d; zSf#wZ&b1;+Bgq92%JTW>{xd?zN~hk244$G%vaF z4>ZYaV}8%z4c%7dQ{B`T4OxnCUBXy+xC9R6#$#@QW@FUT^t4YHQJ z=?~Xs(-i~6pBckRawTSeB(gpn%YoJw&dcowK2$rnP;*)&E5GVM;C9(?A+;FOuBd9! zU9(5$?Y);nk4iG;b=!82Nug0wP++nq#D#45+h#A)pZU2wZ@;D7h;=_DO~uVRFQU)w`Y{<kNXCHpTmOD8S|9FSKx5ZwZi#n4lOO{h zf*)cjVW@KUIUtl)X!TCjWv-98clT007V&JOy>L+x>P>(xB>DLj=9s&cdGBVW?pCAO z&_Jz%eNre`0!2~NiY*t+3OAW!sqKStM6vjs`D57qdxTK)45=oGrFkVRw;um;N7!2C z8{>tMgbwZs?$``jF5q!=8nA{UDfbQ;_$Tw$ad*>5eibeeO#?e2+H8*AK&#appM5q*w?u@r};cq|2I8{ z_|K(tU1qyxgulTXFxc}>-rt=I>M=6X0uC$3WhErqvj&=)Q4jBDv+4qwA}?Tv-~7FA z$apuX5m9p+`*PIrgu&fwMLY-{mkkWxUzu;A5G?Hgdk_lAXTz!;kJmb1k0)B-cS(uK z9h+<@_r$78I&glQ(~MaF2VfIo7bNvt+zkU6p12X6!SYcTFr)smyg{poi~V;}=U=&q zlM9BwE{SXi9VJm{TZfdIP!Zd+XuN~Bw$c*BhmTan!dy^Om)TF};v50Q>Nd@`Uxoef3qm>b-eyjT;51O%*=vIqgi@_*M>(-In%cLbf z-{HciWlFHzKej)}D5VET3_pq(Am9b&ys*^XnbTl;nrMvLSz!Ti+(UK?Gv6Y6>)XbO z$G|lm=ZFtb>!l!YU7T$+ijnlu88KF+J?(T=txJU)8&haj%{#u zK0cPkXXw7nWMSkh#76Z5Yz-}|2N_)!R`yl)jmKpd;bCi%E!f1k!)^EJisHwCbEs2U zl&Wn)$;Uyb|M$W5Q>AZQsLlwUL)s-To6K}0u~jXI`hjL!Z3d+!8*1n7U?F7Uq8cZGNl0zig7P0=P2>rWK~P?0e? zuEYo#7NE-j5g8);>*T-R zQNc8n0~GnW4%tXd;bnve8QWQtV6XZWn0nd0mQLtAfzaLGrxV1v!gJxJ%?>Y3jfedA zZrsbbDk1c+>=PjiK3A?2ABI-tEwgjALm&`=qFkrafJO67PG^Hmt%EGDu4`&*C$La1?}WlQrT^;~d$<)plUs zelEj53#S`iRAD!VhrOUD!@?xpRc@Igy=A0B~bW}YmG z>pcBpxP>$<9>ay(&SN=ONjN{T3j?bqH&Wo&^>H6VI%O9DV}#J8G|#ik&8UCvWjS7D zeLhKuGyDgyRtPG#eD=c9zE|KuGgQBk{kTv5X>O`tbEgFEp>n#slX@kUpL)4<9s_>Y z?(=CSti^hRk!-~3A?fJPWu@Qn<=8uuLFLvWpYy{j;9S;5#N&GnjIDu$-cpz}(jBrE z*5V!bwq+MQz{GvM=*M*W6=}G>+ziDX&&Jy%(V+}gqslUU%Dbm}y|4wSwJ}Npx?9cb zNpKy69>Xb6-Yf9`*HZw%WmSD$aeXy5D--Zceh9d2KLnGrrrl*obg3eL+1j$!a#eNX z(YxhkLWrqm@+4{Y-%t|174L>nudN~VJ3U!HVo(Zi4Fx@B*bH&=(jd%6*^nBY`UG3<~`9{aC*n<6b0J>aO0MkrxCElF)w%BR*hcvt6yGnA*{U6DT z_`MJ%xY85uF{vASI|B7$9tSgi3Le2+PK3hRF77086%l}w&r3Hek|?)-D-lpX>y?O` zw+5xe1?4@-$N$uC4L6T#E)++U0D7Ms@&+z(1`cN<8%9CVigS2_3H`kelNlN2(moST zg^8*4LjK|d1uMk33XvaPiB$1{w&?D3@RFagFxKItr3N6N5iQaLFRel+?`u}vsyEF# zG(UX1R`Csb$KB=wz1p(-9`9%$XjSz7AkU-BsxMRrP3-WTuU}I9>9S!Lk5LN^lxwlb zk^EbyP==E|yde^o#5K$Xhaq$?MU`<`p~@3;+pQ0M{j(1u;-2?SFDfxSo!s5Gj8`a> zO5`}lGXIHywl6byTb9^tT$yCXWj-Y3vs@7v3IP)l$a-q8zoEP^d!c$-sVvVdRkDun(YsNsn`B%ngSf+^s^m>^t|LUt=uha$m&gTm&{D2M+r3$!}E_zV0Hoep$;O zTEj~{M4&ZDo)UO036SAo?fUq1Jx2&qaz*Rfz}wTx+`lZ;oWrB4WN4h`jy~yIKW}E@VX2m!tzzSzXDcw?_ojkv?)$EF@wb9vhddI zSMdVM`du|(ex6l@R!aCqcj>_4hQivk!x4D6vd~rUq82X>0ULrG!F#SfJeGF_Ou?T3 z^<)Ye0NB`@cOE)L7VeZ8qbx%G~I{411Fm|vaN50_+HdDtF^bhtTM=p3GZ zJvfnr+@CTU@BHUx|EofP0LY;pmwWNSM9R8fc5!)4s&pk1@RR#d3ak&SstEe1KLpy0hg?Y6%ZgEfSYt@ zYyR6a@nWAc1q;#CaJC=px^EcitM%D>;9imzZxNh~_&bPaj8wKB@p3v|yfQ3QVCGAa z>ON&MiXQ%+QL`;klXZXKy_|9JjC(14#`|j*(0Jxa7vJMed^2XfL8`Fw1{v`9aa`i1 zPCj0IinQk#D&}Y}`KSL{H6TFBo8)eA;NlsOVGIFx^5D>w?g@uT`sGvx^YqxBl=S|b zaLmc?ajvDq;K>TFNX&Ui2Q`Tt7QHZF*ilw~p^!{5+*d4G3oIHe{>}1?y=;|?l=@EW z&~_O*?)l9gU}68#T=`{F1O(~YaZ&PN1O}*cVo+i-n{gR%5!!~4LY3v99lz4Vg|ZLI+mjP=6j7%BpOpJAKdMQ+Ot-@H@_+cL1yTt+X%DAs29+23ZWjd%BJ zqCXt7%oF5FqE%%;9W? zk$(URj$!)UpoZ{@_`i-uAzK0fH zM^V2MvG46lTISX7fABsadvAySzGfTpU;B|B6>5mN{^y75qhO%VG9J*Xw|fx-BY3S_ zVxO+B!39^zG3fd9k=Mct@#BvJY%XiX%4t?izykN_2}(OhA#f8-hF)k>{etVzM#Tqb zNtarI%D9%I0>kJ7qg3ePlE;zmEu*xxfwz)>R^Fpu;1~c%tumrPwZDn;Z~M_NQY^3? z9}5}HfrSy!=bLkBRA+ARU3o<(C{NT~=9Bu!g9;H@v2|dE^o>G zW-(vdheNkOE~04ue^z+4`gQ46cUi49{tU|EILP(#Wm<*0FMG&CYf;ZwBHOUx-fyrl z>JYx4{#UqM$g-qLzs6GA5_HdL!$>aC^?5_){@?Oenc6~0von-D8jvZcb#*W!pbBA5 z83TaLUG-~NpK*xL3iI{{x%%XMkCOF=y2~`P41;=BH3x^-vBGDtm3n@B3jpHRqM9SL zz7cMfYv+llUHKb`wOKzQDUmXfmY_h{6g1Fn+*~-9|}erTlRchJGgj?MAg+t9N3*qrrPaJEYNAu)uJA z+h;$%mYO}xM;w`Z)ctL1!H|Ck}S)z zdFMWH8~LaHxdVF+Kxhg<-Nte&LX!^$p5QI^aS{oW0ZDc_)7SEmKT8cpFTiR!q7E1@ zjqHzN0cs{6uhwHc&|h9O23U6Ln-Hzvk+d?MSBnMmS&qToma93H6=>P`4D05lSW2NE z3O==)`|}$01G|bg`X4>;9<00Ulis55oVe&YH{_vbZ@m~+-Nt0g1iknYA)Ds&C%#;n z@*9XTV8pzfd_fv!>>dx-su|0NYOJWVEa~lD}AL! zeV;15+oA>Fd+5Kh_#wSk(X%WUaYpBQj5+S287|}h9%1(gr3^gZNi+?ThovbB=ZR7N z4J)Cm6%WkDkI55LfA9C}@=_Hrbaq{pFB3=*F{2}jDmIc%1l6y1*iY2H20J>zsnX36#vQuA|vER+ovvP?)hTW!(&Bj!Se<+)iI*Cj((R zdw%E@fRd3BAn$;oS>_$y)xQVb8hTz%*o-J^9LLL0uEf}m{X1>A(_g=KI;h>UN4eLV z{Z0*eaBM6bjUR;P^(&oZ)$feDG_B4`lGf>~2a#ve-fWEzw$gavFih>-qfhYMAFdOb zGkJOqg?qz(6)44OT#K7TO0upA^QGPBjq8sVOhYbG`J1BoL1%VKEK<4oH=Cm%66UA+ z?#odIaGjihF_-_|AhNGxP1)u(V!9SP+?<&sjKn=Q%@tlZ3h_RxxuMcfa z?eX5kuT)+G0&YB?sp?miWr_jli*pO^_)N*BwTbigSZX7xuQPcP zs(Png_QP>xi(R@N>d|4-vY*dWAqo$Am>qtgPIl*)8Cr!3%G3g;WsZA1?>S=mN_?&` z`Fq&d+ibo$8tgM$ZR(Z+pOOPm1kyb-O#dhSY&8Oz^ zH(1+2`};;=%hlw}%z`FG*7Qk!lmBS2O)Yyt9gN}FB$$k|g?(tUdy7`G}PayhBX-0 znGU|^&^in^^ybvaOs*{WCsXuTx_RW8#^=XL-w>C>Dwm)<)SvEs4% zh%za_`(<*%?JHl+RoCIBKCccY{ap2ozCXi!mWkY zw0fzVVX^IVC*DrFOOV1DpXt1gGurS6n;t-Vjnps7S}*dbH>!yD^t+X%+$AcZ?GZP% z{Ck-TO7MGIj^l!uhn7w%CoxKwTuTvWhP5_A@2}_Ng*>`fDY|*<(qE&EC4G~baK*CU z1N|cB@Hp@8)R7s`sIwlxxLUs!QUlD-(Y3!X#V7>CQM}aZ51Qa{S*$=+N;5v&(cex8 zOM{LYO<^~nXTNW+GPy=`nJzPt+J%ICY1rdWlkv+FkQ152g4Qo)M=_ zx#aahKOBQ>ksx3@B$Lun?q8`4?o=2MoB}9p181# zo=4|=sJDUTAG$yNM+Z6`0#Y750yf>0J#7t`bW}bfO4&yE+EtWq+44w=mz0ZN6WRAV z%;PcbGSrZ-6|;b>1>%tZ|5QxN?0VlYl+SdoSaG7c%!_P0TXI;45qFC$;GdnznL<*i zBdwtyDx6bis!AeM80C=Jo;UW|1|f_&Q$Le@TD9*5A2hUHNi8P?T1(rBIY^=_U~8bD za}mjM`Y}X-bG~U7Zwj03XNWXY245%$rgt~e7HuCaIBM^cwfphBFN>2~*;ArweIH<0 zIlFPiTr1Pg&JccNAZ0<0p+eAPb^g#?Yu2TMxGqHYgF*SLf87*mF|HDT9}cu$d}L0+ z2Wlt(J6+bO(8^(HH{~aN9j8-#O@enyK*L+R#P_yMQ;k5H z$m`DgoMP`l;rs{|-Kw@NJ+iRTv9am~)UcP=E0~Itd?dWRN~e*;j=>nF^se2k!s*`| zOccY~J)#@WV&b~?#X{IW?^DY5@~??In9NddF^6Q%7b#)0egCKVbSaXAo5pBwj{TqD zFeU`g|Lw*57Cq&wu;)?IVwvI#U&djM@j@Rj{&!!slJVqHW;LZ{qx=v{BYty}hb5yR zZG6o_8|=tD;+nJRT=D9f33D<3p}+5K0`V*sE-o#`Su`r z+9X#;Pm9kM!pIk#1FvVcTY;91nCyy?$P&n9@H>JiH0%U!V~Cj57}gP@bfcEn-PF+L z#iI}<>z=}|qWhh?o`3{)H;4~)EGyAOy_@td1NGzH0M9`@wAg%%D#?uBy@RSEEJw;C z((i?;x+_?nw!7Tb(AxXI!gQrwZ!w3@gG9;~;Q~s6D^M_;Tlg?U}1rIp=P09%-+P{Xd)%9wTWywF7j5{i~=i^3FudU z7S{k%WI5;E`B0r5HGa+;SK;!G+$#+Ud_E*~TE1q9M7{S-q!s^TTmf3UvH^CVvO z>TLTu!^x~Gpma1AffB zEIL&g^o~Y{2UgDjwq?oAy--)AA0VB>$P3P1(G_bk;AgYZaL=jJm}f{)(hxa&jXkn> zL+sP5O5R@AA4y}Z0loLEykQ_F&ait@W1x0!7*R%_UKlv~eEwQ*ke-MqmWFP^!>AMv z)htH3uzk6lV|t@?Ke3Q@LXT%G>VdG$EKHDepc_;s=JQG(_I`KuiCxoQ4D zHq%tT-X7O>EX!i`wM)@V;$|mWs*L`2XApN1Cc~*oBS2XQY#SL5IRA(no*#uEw#0|N zoLw_CVv(zr0~fJ zg;%|4)c6}py2&>XX|eKV`4TRJh4%pU7>A))DD|1JPKKn-AN_X3&YDvP#;*xZtY~NJ zwUR!Kjq*sq(VE?GR9spi&#IL0;nC0MQfee=pOCwVgiIpf5;|#^BUGXx?^V)_HTTDI zhKVpnElNGVQ}FZWO5Y?g>^3klL|C5IFNkMhTnhV%28&L;f1byi_m=q+qY z#&sHkicxrqnA;Zf?OWd^NLCYmf|!+7vs9UbctS6Q=o~YJhREYGl$(>6SJ`*TY=|rg zF_?96|ME2D8jvTTDx!o!%an>1k%uI+ve>53p#;BcbIX0MAALsrC{4=ZCIi3utJzT- zJ%s{;P9;Xpz`G5RA~I)wNN=`17hXdCFIW*qwt| zO_LhE!<0sXf=oxJFUy#k6Si@W=O-9xX-ksgE!iJiA`r377+EHdzv*TDCj)=rKt;3M z5?NzaUN#)y@g^u*%G4f#^*(0LFXhAKp=?r zdUEu=`j%MRNIUlDI9!Z#K6B3YN1-qN)|_#!vAm>GG%d%hdwtP=dup&SFK92VGsVRf6V|K5x3WG=qrX+~;=uMCtv_`0oMa%o(2fDTxN{jhv$q>uNWn)Zu zTlHb*E)Du~dtp9KS-{W_Kll&BtaaE75NAF~I#>g<%oyh{m!u*@$@$j!% z1gBVonn~^Ab7PSaBj7E7?LF4EcI%nCE5GDg(p4}*Lmw7&0i)$(EhA3g)yv(${n5pX zdnYWHi4F1cTB{h}?7BhSH4>pa;s}LZYlh{2K&C3Z#${QC=cj;zP|O;jXk!1vmC4;! zpxc|7|64kXmagDi^Wamgg!H);QtKOl-Oq8`MA2p@*SRKiLwP?d=c+Po&*yw6O5m$a zV#`m;SIB-4@GclyJ{xAzfAD$)$XSL4E*y32q9>lPZTCKxw(guQ_(Q*ldrI~ix+$zO zjwZMYZMm%RSYZz^(VXW!_h9bh40OWrbC%w!6d$21bC zGOA1?8}hAQ^R1MZZFcw@acR-w8mAzR@mprmdQui9|E@GoPr(<)*t{gF;dL3N*m0xuad!>RuX&%sPo$u^5=1ZoYoVue{~n8r zq5f2$o|!M#21i@j0}j3Rxm@93+i5uAYP&*jcUgF2bTL%@{?!_M1ZtYX1NpwhsQNkul$|fT&`TM8dQT?B>}vH;i`va z(AtR)i@wABr0Hj`yZ;?U)f;T^N4jso7dT67X!Tgiqp@n0U#H0o+hvHkB6ffFmokRw zZ-t6`VIu4C$T-r)h+lWTq-pOR-8Z@zG&z{*=;^I7kLm)N|JjMF5%YLKlPNJWFEXMJ z>l#)%h&%6Tzd0jR(xOP!gx*L-nw&Rj*PjJTL|kAo_RxUF(>N%7)&)a5^EhwUo5<($ zcY_>;W#H|mDE@kb?WHBX&z2S*6fOLVF^FmF@@iwL=lRW!>_5-LxO_VxZ|^6q-*V{8 zGY$V<)*D?DF(x>`%<(y#7XyQXe}AzuM9i#eq+gi+)+BJ+N3x4|siWB1#&ao7IQ1ic z4bQ&5Tx)(uR)b9{LGN8H7#0?Pxx4nA0=;c(1nv96DV8tfo+j zh{c)~v`^VGG1xm$4Fx{V|(*d{2Rw5RttLf8|FtJ;oXhCB!_u8 zEqbLrjClG*&>!t6{qm}8iq?51<0eba!lWt^v%{`Rp7ms`(T=u>LKkaYqlHvgOuZ>k z=H)5K2zlGtT&DMDAR^}w&beg%1nxAh$%J!MFH{d*i55wu^1i4TOxv~TzY zmM|+%o6r)9iuvR_AiORw<&GgY@zHS)3buF6FF{xMVtI7^2mYVJo;KlXR*_&|nB%B} z>fighBSQs(a>E3&ea?9W%x>l4Dd zAqwd8Ohv}tixDO=ipkN>QtPoF28OLp9M+fY965PsST8FT;8vt6fY4G=fdxih%}^OE za|n@wbmA(@br=>FGC8&gZ>sT%KlL5D$-+C!X%O=Pk0R{7;Q0eOeMtBr># z?%U9Ov7a@7alNH*vn0t6K}{?hGY_A|<~U&iPuC@Tb(ke+#} ztL=UgM?PDFtG?_AtBXGnrEHq@y_8Y0PdE$&9JNjVg(>$JmEV)6j9Kg)m#XKGgV=pr7wdw4Hw(CkeaSqgGR+>PPEU2vnf)%#ml$g*6-~Xv{qNkj+c(nS@o?O z+#tcbw^5uU9|~yo2oO&qFF?&>d}~TSsh(d<(KCG*@XSS{clkBswdPjv3i=rAyn=jY z6LuX6?;9+((ppRc{IA1=s|^C_21jqlHi0=JpZ?UCnBBU?;F%SFcB;q7W@UxCq>~^xhp+#N;qLsu;=Ffo}xG_b5*! zNb`=oeiF1YXVvjFSJbeuRsIym`Q;N6hI+Ob?P4-@W|wen9IUO^{coIgEZWRm;@s|6lv`yFn!gU z(DF5Ykg)Kq&<4W|1YTy$gz7WYdYv))ruJgizhjmDoTO-yP1Q%nSI`@se^S6kZkU1> zOmaR}5|Ug`_s7NJGFGji^RN0l8aNI)pR2WknjAaozGyltJB{m)dU6|cDk|}BCzQxm z@u@O_1WX(J(E~GC{zSYoihjx}6;Nd3;F;-SxI~GKO zUOtQ@QzIO`i6lBn8jwZQq+9X}U~yz56qusihj%WoqHJQ%J>YZTe00o4?5G}3Wy3up zY(jovtRctC(Ixt+q2Z<|xzDd5d^YZBdJ{JHb5B|-PL2GK$`eb#w`FR) zgL#pVdi0OLs>${OO&L<5>o!d9Hd7@nugqo~nJWQ}cVCb>=5E7+go43ka&kmNE>A?P zRmUr_eI9#z#bnOOoYCKsRr-J^8;#ID`0q2)B70!Z+;#{W5Z8{0{X^!L=~(Y5?~n5A zH$wt3z(d-64%S$^-M(O+IgE=wNVkKIl&LIuAuO82R$GBr4bwh%pD}P6`7~$Apma{~ zmS!8>-OBBK z(-p5$B()hW$S=y0VkL_K8vb^NoaTXu?07j&7#NJzFG4Jn!CY)9)Qd8q4a1R8ma7D1B#|CjIFlIHU)pRr^tA3Lh5eSo5@FC zx+q-aJM9j47aYW=s>%=X{0;thHu`KuQDm zZrivQ_<+8Yhw)aFm56zRmNqW zHw4S1=KEwc_yt`%p!z~(pmxAK0gPDuF`g+oArIz-7u7|FPLmQbGjvRpYI4An$-_zW zL|cPj?RC-<_d)4}J|W3UZ6;T^a(ULGSaQC14|87Ztia#-_!ko23)BMT{)_Fg3u#+sU2w%+L$1x9aGem3;D1{pktsL z*tF~UJHJrya0oyx><%i~3fpC`215YVsLQZ=4kBXk2mAw|-Y^uT_WAfyq{Cvewc}IG zhp-yYa4P__!Z#yE6iZt$zNB%M3dC9v4H^jDEg^I9!5Fo9-$P2#r)?Oqlfz5`^yAY6 zwMF6j_aNVD6RK&yiy`E&EieitZxw!eo2)L?y9s|%e(Ryzc^61(@&C@7F!1Oka5R_? zh4LhCY3j~Q3Lg0UC6~FYMi`6~3>y}IbI?=Uldw9iHAwK}RZ@z4?=&LIn{UnbDl>@F zA@eEB*X!mo76xPHjRWd~PUOx}Z z-Ph;Mx-6a_L?Ak5F}7Z?U$W!Bg2^nWiUq2KM0^4il9HrRl9Jc!o;X`)r!@SPe07Ni z9s~DPi7#I}&VjX0=MJZmKOn|^H@yEdkh!(3Mj#(J5C0%&U`XlqU*%?>O6*zmtYi*w z`-5NI8SGW?pnV~vlu(-&dY<=7plaTT`W8E#qH+1AtQ)Qo1IvCvxaYI~-Z@_PE*E*X z5y|rvDPMZcHe(X*?0~-EXuG9fki>sLF|p z!eqB?j1m+dAHPF7)@8ze*OFXZFsiQioY+MFlw3r*%W6BB=0QYMRv18x0bEH9L>#(&y_=D= z-VsH_kGpG^qlEq)&;z$i#zaHkfsIt4u%&k6`g|q09sf6@*?BkZ=hC->%8B=z-d9t` zu#x@#-;t$d%m4&v^0D0Wq;KFHL$?!8(^@@)(-GAW9M9T48+D>IxWy^@?ZzcCYc_n$ zRNJ{+!|ZQvYwj1cpf>(c(+ZaU$(XFXOggf z3T=K63(e}5-|C9|C;C6nn`Mb^wbL8=a_wu%V2-Wl3SK#mAJtGx7-X8%H)*1UA1V=? zi#;-$VaM@Z>9|5(cFs0U&D9pR=#EYk^ExYp(;?%F_e)Q`mp~GIqZz`sRilHc*VYj} z&xZ-`KwsqlUeD!`#Q&}*&52X3`eV}kZ(14`C+h9FrbEZip3^Mk;PT|sCf-dp2L3f! z+-A_)%^`x(y@?Vtss1|Vw$b)!670iUB8-3)OEAS&Qr}uYE*zZePr)naGoKIuCL4xf zyc+~jPXXLBQo00vycq?2mMMHp&1}D)vU3BsK0fac_v%D(m%*5u&Ya?}Hj{#hJ=yI6 zha%~<7m-($*5wl^#_bpH0((@9$rUC3pjfK>?;s2tg#R0alZj|;&)ToMoDSmu=FO;o zD&xO-)=hY7$$w(_{}cT$^=zf1{|5EHFObUpZ~86fkN*5+TdwyG2jLE-;~RAO@76H*t%hivx{1>VS6qm12VVI#j(is1QjGvj>J zE*|>L=*TxC`kam=84*gSwL^99Ip40}*Hm#eyKDdKxn*2^T+VqVwR;WrH2Eaz^dZ!% z`?}T|eBY-*l{J`gvgAj*LAm`2gpLJkVkK*%dFf1hhV_q0poHD$f$jxu*g0fHlqs2I zd%AvGf6K&eP#E+d7su|@F?CUE{=?xr{YU(=IOLOg9Lx|@4bkz;9}cdWd?$$A_V#hg zXWh7wUqh*}Ya&`jW?XBz-RWQESN)=I?yf$(ei-fYhSB^}Rx``893C}{B!OmUP|WM( zflcg90|R-qdb^QfGd@8GsI|C-=Bn;_XiyUIDQXrqv*mlR-w$V&5v>DbH8BuMerO7v ze?}hN0ZgNA2=?UkevRXEjTO77%2eU%x&Kugc10wtHM22@sfU@DV&P`R6OrXg2pZo} z)<|lGALkf09!&Q?B|@W~)G?}T zpgcyf+rKx-&98#-9Pz(seN(J^C;p>Ww}PH;l+TGeei`&s@@(Dx9f*oPf7_fjiND?W z{W0x47&qOQ6RnkOi>9~EHR~RuuM{T5f8noMZcbYu*}am(8hiQn{rNc0v-psZ!y1O9 z;>E7#LdB(E=fHohQ%W9j;`zJ-eLfarftOccLL-OKLO^#`va0EvIrxb@k}u23>+QKW zp60Ct(Er=_pEwCxy6$DGmE3nX&5YkpoQ&Vpyww>@nQC7D4EeT$ah|P+ZlU(shVStg zIZm|P*31_G&*LB8OEIbqquI2IXZOB2vG8;Gfn5OM*=~Ni)6qzO>Go?=UAB2P(t1`F zx>@u&-2K3ITwi-U@lVQnia^i(CAS-O#(oe9R?{cwMLElCq<-a4mh`tCOJkL8&AvT6p#O&N30E&e*0+jj#+HfT^>;6-_~qAGWt=;A z!t}^@%9}6)EO)IW@}~D=K(ULqyZ`F{#-NWwy1~*6ykPj^@26vnMfOvE!LoR-`5ANsyn_&5+_Q$IX{S@sx~Qd6!)&gN$i0_4%2FfV`QQhwrwMO_I;}PUX4< z21x8BFDs>~vTG?YGX_56H^RcX>1^R@9(nzdHW>q6+rbV-_9~`N*jBjqVx&Xz1$0?w znK!N=tb#~D_;0=~RQo@VipzYK2K9LbZMn!dkvISeR@6RJWg@O!)NO)n(GH}ODKVR0tU?<&0U~*mHD%&t#5pO z#y7?DK1&M|bTH$bB;Ml54H-!I^p0?W$#2Gx87#cc7=h!IJGz>izMEeYNAWI!@d)sT zU^0Ftp$uOG2em!R)skc)lq-21lDo+CYlE6)nJ~%IoI0 z?tR@mSu439U{In2eVZ|3mZeItH|X@7;Q3YM`@?7FsI0A}wXkLCU-u)D&I9dtPc0Y2 z$L6c@F`FE%?FULk=uW}lZC-lk#OqaWEdpVwuON1hYand{acCJPDd!iI#2Z{nfltWl zLED#=D|g)oc+$BMAp^bO+qiggK3$q%5G~TGDYLjK&?Gj!4uMHa5D&^j$~E>X8>Nij z7x=HdG#e#{#)Jt%uHCJ7W$FPOd_63PrkQU++$9{}x>+}E02A%Sapyn9rm}t6i%`0q zQGCEcTKfG!6-=J#rn@-+HdYvpK^H(f@Q=a@`54b^-erWY-rPCu z4#;a#cODMX(`9=fVbMfSe&3dQ?AO~Hwf_?t$Lp|YiT}LLwp`?HvkC6j~ytw-NWQ^ASDRv{c=iRQ0E5E?v$i_sFhj_>xxCw4A$oC zli#w<4^`O_hb5byc9!WT$m=~jyjFS3HF^-jS^+=A(bjU9aES*YEVQ619t=0Kwl1Sq z7hd32;IMZ^3R+)+BW(;fw1xC$Cw%EyI+GIPC>-qzbklcR3FXgof{7>$ZlhZ4wd#RB zU{+@y3f_hWp@Rc2QDk9~GeozHPjqy*h1(KqEp`6!sN3y!Bd3-ABpCl_Ww+90c~~c1R4R^NgM5+lapLfMdE?z>v$>=uWz^8Fl?Oi{>b+aW1^B6 zwo|_ts_XBZR1S2n|4_}BYbh$BQ}H=(G%3L>l9-=1eE})l>qR$+xMS{BH!KS!;~>E3 zbHAF_zfPZ?J@I3)aA>Ni>c+`m;I(GnmV`?VTBPTkrO62wfj&@Yi~cv~N4! z445kTwY%I42> zi8?E%o(N@k<+Y+4l%8f9*TgCE&cTpy@!}huaUmaL{2?q#$W{;XL;tK0=g4kJ{C;^V zl|*EA6IdI8YOs#a(gvdrQj)N)$8N6WmU# zT;Wz9a4Q8AdU1U$-IGFkV?9u?^TKrhv@HQSGXeJBZ1EO$=jRAx?8AyKAoBPrYf#i z9wvy!mU%0__qn**YE}sn^3Z!*ZM3Z?+|n06mk|z+!=(|^S59LM&ABT;`=g_r#%Vy; zD_tjBz~Z}nIdB;$9AFDtn zG^{;^MN{b+B2->wb7TEOYt^S8idj&oSa|J zv7t#;ur$>wF1(g7+K5_?FnZ{=v@j~I?Tgpx+>_?O=1C>brI-DJ6@_p*$nyw@7#MmP zt97 za{n^mha&p@R4X|RmCpton=@TcCabg=)a7xgyu(raYfN`U+`fRypxc3gh~Kt|>PUdW zXVeqh>|6W0ECJ_;!$Av}qO4-8k@~L5_ex@iFUL@7wj(qRiRj352 zyTftBgPdx-j?20vocd_}?j9f~4tNn|IyP>-rkVne8oetC%|-?6^2kIERyJIV zU|1cPK}?LwTejZljElv%=m)!5i?MGDDVrfxAnFo(2!PP@a~h}3I!l2aE5m5Mv5=cx zuiMjC0ewj#b`6bDZ=%EQuTNAxFx<+T%Qj^%`@^!xR4|<>x|jhSa9MJ(d4X)Lmvl@gz8zi z_*Wl$sWIM+yMf3iHxB3u-Yg8=cE zSIWBoqS<#7pH)vAEuo+Jlg@MN+`jVf>Ftv5l)%^Y4tGbYw*iT{b^1ba$5NbK z%fk}`!*PE*qK>W)!r64I`Zn_V*vWXvkHawz12uweqU4cZGcdY9*ZprW->X^4K#{_? zJDy|f%{A{jR$&(WUHrZ&oC|fLssxM>8nPAl=UZ!!-N1P5GVRa)(=b_=_}QxdQ(bSn z=kZS%r&Z;VYW3iIuIpH##>T8nf{-or(s>rI>n%f0QC{{rU1ws~zux6d#Q(PZ@g@Eo z>}1PaEt%v}>qXy=VP!U}SplnaT2CHT-*OU7g`;94s6mEKWXvrh92n`)g&!<-C`%#J5jYD}OUEW?(AI-vv@HlFZJEtQ&}aoT|BE$W~*V zhy8`^B;#7bOakx-Nl(khbGYh#Pr;8ypq3tnfPB-N+9^9N<~f<{lR4jLl8urpNfgZ1 z8IbG4t1TX+x){0+4~0p)2R*xT{8Rf3k0^GNY;KPp;yuGAz#Q;R_1$Af+m6om8@nIX#e?EV4-Y1{_C>=IxNkhc_OL_t6;j}+IyaR1k;XnP zmlx@GJcHp!Do5h(A-1EnzvGfMsnDrhQOYc2CYe>~%0!(Ph^_hUvS7Ml6ne2o=s5W0 z_#WYM9JQcYRS^0%8W@My>?MUz|Ih`*$~($|c%r99 zi4U~v1~5_MzX5YRFK z`fq2n0xNonWwhR+9_tHdO1d#N{tp9}>BytMc6VP+3{wWDR)?C_u*WX5Ikq87;Z z_6ADa&+Jb*!wryr1P}q0j0aZn4OU-Q${i7%ywFEm?eW)GdD`G?nQ6JW z9KX;vi!P5d@jFgqe;DL{{#UttPhfFP#iZ8WEsI$L-@_^!nDfAI)nOmpUHr?JMzchV zxt{p`ggi;~-TK$2QiGXxKAfHnBRM|{KG!VYyJd7q&da0@sNch17gj2V)rMlayFy$^ zP52viNfR_c>ySwq5h zF!7idx-pSKz2-2L&YSPKi;Jp!MxV|f9&#}^ytOweg+vHK*!BUQ;csU#cJd!i2vv%H zxRLEMR=)h@{2r6KnX&C9bq_ncvYJ=90ewcU<*0x{tDYC-0GT?_NR}%cQ#0+!RKm^q zkE+Z@3}Scn7NBxTeyw)!p9~&EE*fiX?rd~r0zY5(1%IS~<2Z!K{Dl(r@TC2B5oO~c zy)?HpC$z*72}CWyV&_)&iSAILOVl>Z&iBu>b|sdzd=U@EUS5J9s32-=dR6SUYI#Xq z^_3lZwWf8N`c)Y3FlulME*kUwJFxc$-T%S}G@e*?LF6w~66O%GcB5)A2yCNJade|FJ&n5mW%an`~>lpzTfGL8BU(!Q0+s6@}9?<~^% z9y`v>EUW?7e-dcW@TFYr5oD!NAt(FU6D9OTV6V6u<;c8Zp2kGV-DI)*n5Nb2kA@Nt zf!`#2y6gT36p8OnkO3L#23FKcj&Hi}3;38L@Et6+YfBufs3@hSlFJ;iyP>X4aO z;J98iPxrlJ$imJ3sM2=uo3w4=?s9W6#=)tBBcw~Rz)y&ca<&zs`2E}&eyJt_YAh3&f1wipP7zco*0V$saFk=p_8c& znV01mYv0?@K)=0s+E!uE2i`~T?H$%|As!dtZIy=6VNPgEd<$79RK&58pdCV0bq@33 zq5_!gk0$CNb5_fH4E$Y5DeslqeM&1*lO_FltWEoaWFoa5ZPA;8^Czop{xMj#w!d;uV+%K0k1C$yfsFdOs)&+78Lmxr{I{2Xf!LG7XQjGJjp(S=DTDG8j+CJtGf1 zttDjo8@Xg}GGl#k$EBd6*Jk(w?|kWio(ia~t6goXTD>~kL}Cbb-6UYS3u%3|p#80f zb@wAtl8SZs+UfV9W~DBNG|+XDUKQ~3RI7BHNh6Bs3$hoH@;+b`%6zt5J3107tq~P5 z7KDA7e43FsdV}>wOo>{dTZsWyXfRUKOQE&P%O<$syE)F~D}A$xNtVtO^jG1(n(ehm zVQ2}Xxj2~{;O~L1JUQ7>gD1O(~Ts%CD3fRnAS*#Ck#f+0cAKr<=(*P_SBTC!h$NCg?; z&mO#85r=BiwF&JUVq(U<;*NhS7%tXY_#)wZyQUbYw!hsJs$#<8`HoA20gpk%Pc-ro ziJ_E`IFUyMQyG(B2}%fMSpG3kxi>iIX7j@gonobA=v9q$IEv*0U9Sq(UQ)iQI9)}Yz({$Xf&SgyZaFNL><}&iTa$g^dwYEv z6@1lic5~eT1;;6MPvt<7POtI?)$PD~utGLiE(Z#A24@u)17><~?P_R>oezKcys_vu z53##(dJQU3y(}XtXNw#yzCENly*e2*Ol*3e zDcdkhx1D9xNC!xd%|y%b>ieWby0vK(gbh$v$!Wl@y*k241fun)X)L}jXhYus;JHyN zaX1j73Agkd?@X>oB-c&XnULgh1A1eLTbIAG$=P6WptR#^?PKBxz?Y#To z%!Dg(F%TcIxY{h!@N}cPXMPS#pd=$#zA8JgQbt>4%_UB z61QkaOxQkkgAAB286wsWy9ursLvd24Jm}ty(kFlX{myTW2)cu^FpiR?&4o~vX;{4* z%xg|!m3sLfnw7jreWLzke)!>|h6b42(9wNuv^@2Iol^Y+w3R6lyiDc&U67+csao@J zkd`v$0tWoZAFsPUQDIhwcsk{VIe!;2oW`}qVIM8EG*aa;fSGOLpFpE`4gFX)9aj0n)t6@v~p^U@z z^AO7lhoWL+omRxwlz$qCqmS6%kWOLIvbnE}#o#%;W#%`Yp6!Uj_@(9-+TOUB;{l%r zr&Zv`VpiX^I8qaXD?v6e?zJ}XO10r%9I}%W=2L@?v9(-dz}7oBwc_bkAa`gc!EE(Q zMk9wjja;<;tL!N}hud3TBm-tObyv&>AVF-(QKC3XWs0a*X>X5Mfu=BX?1{j)PO00Y z@j?=wDQQtH?&W%(^c7YJRw6n!9}fqPrK-(DinR2HulJv;I%W0KqrWf`nZx#@)ZfF( zQqq~26ZyserftQEVnof059hmNp>gJ!L)33N721)!SMeR|d)Y^*ISNGVa9y^u4NoY& zY}Ip|%N)M<3u}%mqW}wA>{q-}Wxfs(dXwzeqcFpj#opE)m~8`M+Lyz+^`7{yxz5%m zWV3t*RZeCxA;k(ke#BvrV#)Fs|{by_@z2;BCesg$6@w<#J&#^e$(f@4y5172OH zlu{Bw8flh*)9mX$pc|pmn`q{febu*oDRr0gIMrB zOC1X^EDE_kFc9Nh0?gu$_WwI`B{$*`uEspI7bp@0{&~ ztVXaZ%vExE+vj1XGY}OGJ1`2Uhba2tvGbSdaYZYwlimFL1`?~Y2>%jMiTU+|@?@R3 z;5U__KL|xT#dXnQr#LqCXjx}#sZFZ5P>|t1IV*8PlQl9mt!!1en z4@QaK+x(~Mbj;@YASe7<$-O>cM#l*Q8_)Fc#Nz2GH<#`#>}{ErH>`nHvi?u3nl9&H zBRgrZXY&c~ZFOW;B+xeY8S=q-#4A!Ap1g>Djb%SM7hojj3!*v{)9(_qN*jO73skAi z?<+5Hrr-u0{#JaA63yo{{kTb8)%M2%b(aud8k}4fb|2mR+mZEV8Njj8JJLzJoXu$a zwrfuEn;4W)c}XFN`O7y`=TG6*V1V&l5cON4q%k~%%8An}dBI}f%aXA7`mQj{ z!6h=f(pt=SHhMp&39CyqV*fS7#aNZs_cq=6W^~F?K}A(Zomy*r@qo+K%h7ZY8h zV{<{_lwuTdb|Sp)4Z7khLUtOk)WjsGdX)xk;UYL&uL_ltA_g3^5d^k4f+X{>$MCUU ztHfnSSg!k*ClpuJ9?d356s3j! zits;-5;kn~^1l=gR|K!Yq8wfF2HcJxg%23}J#I21=3qqbUYLi>0L-cJ2bx8Ej1(|i z&&9Ox>u!&BlzF53-RQ3~l(A*AD|D5rE{ttK=;L#q-*xSGNZ9q%s~#RyPabL<@iJdY zCH;)B+uE#n|(JCzryhU#UmnJ{X)qVBCT=syjTl#Iqdu^wO*b3BbNCm-0pJ zk4LbeB>;#qTAl5DiW6&gs?Gko$p-&_069U%zV)?qlHYd+r9l1jPZz zPlDm1jD^6MJB#^XdbD|S88vd0!luH&CI*R33k#KAz52*G2o}GB>kzseqPv$vhmY#+ z?;O~9eCfg^C}!rO--pzF7W+U#PJ-?-5G{qayLaqTq1SmRSLnv;82&V&KXc)3XFUX` z6`|1Jbq3JoeGHP#UB{_YrBOxAeguE#F8GV+JMqAvx49sgUJC_OGVB_s5SQ*tEPMtl zr!Pgt84gL&A?}Vp1#62C;9n7fCvOS}ktxh3tc(!IQGjRr7K?h^P0kn=hJLX4crW3j zJvN^NtD8c>tWc6LfHeiAbT`HYOZc2C`{2Dq+~_^T%_X&#iHwehkTwUb3eKCp;+MOU zscb4yq9ha=7j$O=ElL(GS|TRii_;s>Nrto;5ZHOA;+J$lH(HxFZieff2)Mk7fh)A+ zvSrg|?Cd|U^uYPo^+#F&tV5U!?tGxtKo|u6MG6;DWr`Uv;c#CuD}$zR11*4{vv3mw z3o~v2-5O1VaGY~NcUuSGeuo0cl$gKYrvD1pLq}x=gw(_@@z0(8yv{yoLWX$?f2ojp zz31x)n0)}WHD3+?< z&OhBh(S=yPe0k->ak?ks8pM5=BMNj6KmmIAunFc){AYrKYzqY8RIZ(e)eMz;{NwyE zUBr%kT7aa7J2I{h#!5L0I;88SxxmS8D9AWxaTiIWr&V#Ov=69{HG2$}MoVQkloa{H z;r0zQNc?bpbNv$kVcr_yB5*Hje-Q!V!oIXP)DEhEgFz>n^ag95#(!lB^1Z!v=1mzq*gJV08 z)E(1OSQ&7>3l=P_LVwa9W8WxJO^>ua(wcz@r6+kG;HLr2tzW+uFNOJ)|D1&zI*x(- z?;LP=8oZkP7mMDE8RH<@sn}yNk5uT9Cmcij;aDFhVU@=IQDJBM$@zyd6H=TiDhMA< zB8~;jtMNj59Lim-i+WV#3`3j=#>LnID;Bz4BR@!!Hmwwfo2t} z_rNM;Bi8N(KqfFWjsu@5ngA>b{?T>SdOQoV!^((ODcQ4TSJ!xFvF0x#uzVc)%<(gk ze{ufFe~6DHSQ9!XHji4#LX*kR4?k4bhTQcQ8XAVN({6}`bd`&;1NJ<5dWaO zTTOp_WFYC%WnakW_zzvvOWyAUyYq34 z9r@`<^li9qj-#TAeHrr=6js6Kf4udV*!aI3{Sj~w|JU*U$#tR(1wULb_LSv1iH6(o z^mhOLX75#6rd?KwVYfXALL42Fr^pQ6Ti_kHr?w9dTA3dd`c^9F&Vg_CL-P6OpQ~9LeVfx?JO!R)_|vgl>?iD? zF9wrD2In3A>2P(M#GTf$aJ{t&?xQIDq`;j1;wf-02TKb+Pusq3@4RDL8m4QVOl_&HmPld^*%^M+{ z?k}_F{2@R8{F6NL$ioQ8J4&y=*8Mf<0Moz)5ROq}PoMG>cpW@+Kp2yPvr(aZc@@0& z!S3|%@Nf;x&WEC%QI=8g&)v=xc5-(LEgiUnI0KY5f1NmvUC*r*Zxm+j*|S@N+)^vw z8dfz&j~~T@dXxs$j)FU+ED$a)M;rpiyVwJq>6-;QLZ9(`fm;q=^%R29MN>)$Zy67P zaZ&of4b^sB&xl>@Hqq*aZAC_H)j3_he1*D_V!Q;-Gj}j--ntp?q0Y(!aEU{u%24c% z&XgJPCnBSDXY*RbIwAh&La=xG^l1}E0psBO(}fu0qYQ<6mjVSMELbUA$i$pDbodZl z`JB=h3(o&s&{3~mZ>sw+R^|@qhK(Ak+os;YFRd}?mM1h>QtS-J4gpwEy!7G=y7QdM zEGm!=Q6LZ8(LEO9Akadg4PIgyqkub981$HmHYN*y&6-u%`EpE^a|_@ait${iOkg^N z{(blD_d;{%d*Pz({`>ETOPXOSz&!$3&py*hUwl4<`!UiVU!ZBlv}yBZ6*%wPyHDHU z{8M?yIljBj-FUHX3zOBp5>X%=u#ajWWv&gnhT>cuKp^QAD>0X`7iAz`*eFze_Sw!t z*F+A7)SV~kAati-`5c6xOk8`kblH#lN&ge@Voy4oH}6l4p+k3N6!_9@9bFXtFz$QY zWvI1Sv7$<+jG03m(N)iJ(C6-5d*l`LZPLU^O5YhWrpH{vWft0?5`->=zWeq&#Htve zeIWj6DaH6G10X2Q4ng@oDEsIxk8x!d!u1<1d8kb2ftM;OHpau14`X3;$C{(kbsApu zxNeISErJ)l85*yKRsoC^!})L9_DOjHZhlfhLD&uMzi7qv3oKq}aYiKs1-`w3$M3)U zUKTB0q;8wGLODZW@sh<$lvb>@8OP7Q(%OV>k7!{)w|aD&MNa=$DP~brXY*GiC^tvafV|MSht)307ef;ZCfuD&S6HF3F=f zKa6jY9&y3coqx8Zkc8OI=?At%hODhe0wS8ZV0dJaXi)x+0|8z}mPjDX8H z1{NoMbcWeCtryAXBEYlu9oUBtffKr(`OIPLDy~0TTrrl>8myH%ch`aIsRq~u*H3p` z(ZY|`Xr#+hh*i_`jh?DVTnxISHPO3+2V+ygDCzKY2UsaRfL@vP$GAmQ_R<9y`@%q1+^8T>n(;693)tal>7XD_5;RjF~XZ&3xczhq{0~isvoU-@bqROWuDJ8T!OW4)f0h zl)CiKteb#;`~3Hnxj8qZbxb2}nFV1BZhQRS>iG}CcL>o^434*e{*;Wh3&>gcw$K!@ zV@?foH1tog4Zp7S5^qIvRuA%O<*JK-Rcs{+mz*;w+*#~g!M!|j)}JjjMdq6RnhQbr zL`D$*LHfJS^#?MtOO{!8{ppGGW7ndm>}Egz_|5gj7hed!)p_Li)iJy~9zAwMze}j} zcoFVh>CQVmJY0Kv`#*m%cSt_;B6l79j~niG;Jcfa+1#+_<-fExe;x4~m_CLjg2MFh zUn)DOP#7_Cq);)!3RHgZJCkl}d5t^%Eg!!v`LB-d@1H+?;TnsU5{>R}sCWv*Kfi(a z-J{7&39@iMhHo=kV-UYRh)WOurRy=evgKl;;Gg*O;@{$po=bZ4=*i2dVAgIp+GCqc zlmYnG779?2@Ab^^yWY0r#eYk<-DLvlAo<2WcivpkgXu42E}Gb9EQ}0jYtX=Vv}gci zE6zo2BoMx7R~t_dKKsZN2%qC``9v#hjNG8HLm025W2cU~xp5c9X~*x}&K43KzWE4Z z%eIrj_mANFmqQI6{-*?d&Gg5+ob|@6kC1t0NY8oybuVH~Flp7>;QI%gAJ~h4#1ijd z0?*y*@h`+LaVz^Ws6^&q7v+~-rC<>#vBbUNzrt>@x3)5ZTT7aK~#?M`>RH!hHfu?}*Q?O*)wAU#-zKEU0bVsyf z`wm&UbO{s+$K>e_Pn$MuMZbyihtJpv>~9hXRhiCVXEzlmv>Nc@UoYZ^{h*?Q!l_ML zHo+y#M(o6W$ZeEkpdgI`NCvreFW3e!F2=9m?)wxFWP1~L@W)_Rd2$GMiKK+!+xpy9 zY!yRV<{*`yJKQmzk}k6RFW{kuUDX(;Q`Nbi@PYk^#WvjJj46>&3Ty=D+)Ygz0(TDe zKI}jS&O(!5*EV)NfByMs2;KjZ`t|DR%R=oswPEQ{(uJV-MZGZW;^z4I?Hq%48AzFl z^TXZdBz;;0*rfQN1px3$-0QAry(h4v@n*y}0c9O%kk&eA#=<`bYl=`P>_{J^zZeKb z37>9${KzUk7^hwT3+(3wj3))=n0@5f82cq&!fjWY|G-I9iz)P*_$U4O5_itW%Ix1~ zt01*1-00XI#amOpo0JY=1|i&;9+(&lg|3k3!p$Zf7uHe=m+tHp3(cG>elbsqKNQib zfEFTLE5tuLr)$wOe)f|Bbh}0==;i!#UfhKGj5on&Ng)0CPZTX|jCFx+&us%4vU>K3 zbnnH#TW;sde^_`=!@9$|riC2GVAeic=86Q}f{rN2HYI^QGH_XwxUiMH@JaW6`~r=_ zKE44|gfkz9dOW^tJ7x^pKw%bSB5Zd>q1iI?I zta+|yu0L9<(6Yyy2y@!hsfe$WS*pS{ou9; zFL5jTGN|N*zkU05Rao&Oof`o>WNq%bY@_in#4quIb#tiXh3}TzOkg8{=eAzFj>hl0 zTh*Hh!Vlu#^ZxH84b2DATj05SL;MTzOWex73@Rl=|GEw%3aATAzW<{V?qOv-(~z~N zrXgMfBgX~fO2t6tJBrH@o!hm(%{#Zn!R(UO%H|1*8umIf{3PR}#sUn&58|I>;nD~x ziYGz(ORV)r!Q>dYvH9}L|5NLB-y-HZbkL21e~N!0eu@7bdWK2~@gKKlU?YL&m#uh( zuDbUia~s0vhMn!O1fUf&T@YU{|6`ohF;Iw5hC?d?m%lp8hMVK^@8Q%TG82FO(Hf#bgZrdL_3ASH*Xe-eEO^Y0v_ zju1G0`~2ZXB;Jtt002M$NklSdeJNtH8f$Fv$NCi~qax!%PtWe@p*k$9{uY zK^@>~9Ey2(X0XXw;`TJ*BP!jF&p(gE4gJj$KZSr)FmQAUjo(#-aQs9m{wVaNA9$vdr%q}-iq@@K>$&^$H~3d!KXIu@jZ?*n zKB_pZ0Q|E}2DerhG|1BpOyL3rS?rF{3Zh-%S|EJPvBvT+cR6rz0(exAaM!MOMVs`W zf4v14*`%uo3BW)5>{EoWT_Q^x!%@E;bzIsOe=IOP@`pumF~hmQefF|mFc)3?Yncar zsLr2T(_{eO;)L-I>ea8OJNqY0nkZE&RhE*)9Yu%(pYD0Aiyy^5x@;98Y>@P2hok@; z_%1$B=;EFgJeAE@EOqJKz5(JQ@z22P-uX`f0XETR6BTD{(oLT6yJg-?0Q@9Lp#)26 zlphqvy2>VJ{JhonFd_QmglLnt4-R}QFy##UB~+%Mz5)Vxr%i`|-EQlkWy}ZL(*Xv+ zzta5EimL=HaWr}GA72qdaOkMhSAw9Whg*Ogl};8u@z1>Nk{!Rh_EDb~sRVJF#5oF_ z`>1q{mjd8(MbQnCn=CLr0`q8?F4hRg!Y3>$V^qkE@`UhDg#oQZ;w4mOn^c4tPv_wS^sOjmm;lAE|mEMe=kPH_=}oGSY`Ko5M|LofbO!=wQCv{nfXHS3v5Qh!AV zZb^Xt40e6(Y2O3jE6t-{asF=~+E3!TR_$6EN9$VVUqN7?R|DV9zq!ZhD-QwVznSoF z_4|)v;TDcR2>;g7Ul9KW|H=d$o*4oKF#X7rVUt%FK5CA9)^Jy$EjGLa3Vd)NEoOm{ zRVW*=?V6p4;8M6Aiy!+e;QHsLg#TynIsmgMn)c9ZLhmi1_YTqtAWaYu1yqU%s9^i- zy;m}Xjo26>JPd#$tG^K+Z;Kdde`?_d8V_(q}3Y46~lhkr2t zX*tcR7%zoK{Q=@F|7ENc@><@bGJqRVqT|1I@NE@J3%H06{_kxk&;LpPl}fdp?DLKN zx9bo4&Qt)wz6mq7>ks3AGSNC@IFu6{H{wL;ngouzG6ZB$x47KQTOcUEqSCCH90;@i zX(jmau756$G5-1x8jIrk%o|Wh8LN~q#5w6MRLU@JNOQ#HxEJ0wINs5%zw6=ogY|}? zzds|sB^O?uzFIMFvC*QQ54`Z>YyA((A0$1VeiGhj=BlR&_x&I3(r;z7>mQ9j`{jTC zP+Z=O{y7hry!7YCKM)cN`cuR?_~XIOf6eEQ$#jmUpy3c^zS-N}@;WEYRK#;Ak$~&3 z5l7fop%HF7&_W8yOtJ-Vea-fiaEd7*@*Eh@g$d`*EdwUO;2VJ+ z8R|VS7o;GU-Sj0-r#+``V^Pqw+qT|$(q8J(Q^Lvj_1~!g`Od9K{V!Fzq`Cvq@kk)} zbkAhl5^2yA2)=?z!1344#(@9W_CF{nNS=B2S$(OBv@>53X|U-z<$7!e+#Ye$lyV^c z%lhwYecaOj&G1jqPvu8BxW+02mvM}1!cy{!0Y_V_qo#jb{9P{NL23gA1yx~_@8O9Q(+AK-xG?y zYf2)483P?WnVeg}rw0rsx^(6r6QP+^OfQ`rbBJ#JV~jja7#CVPn#IcEp6yxA+ zebiIRH)aGaD$>Gq(pacCV`A&FM+XXn`|jJj53V~38Y>DS z$PQkVvT-y1zV4Z`$o=L| zj+5`iJ@b?@c^0lcvOZ0>-+vdFb3(T-X<=2scpJA1p8?v>AWlW$qD5}RRW$wQJUy^~ zFM`v97N9Jl>!>|&MMIqD$(>tW(G&qLV}pN#4AEUfpLMjCvKB`IR-xC3eoLbP*9LU@%_Zyf41!`;^5Gi z?E#oGXUm!$adl`F*C%=kAr@b#ZDHyAIZMDY@)i}47bJLBlTmhKgckdBe)5{xlmiWV*c>lDzInv$U61h;?f=sFuj|w^AG)HgWMEz z|1Ft&ilV=sp1G&)$-kigZf~3ufBzGf{`~m27%;5Bgg$krB))t(#YmMrVHvaAkSDkY zsSb0syiPloFp?f#YQ7%Lvp9uSh{L9M!8%b$AH^m2!K#G`?=kt|`|+>z;itd&oPV5Z zoVlEVe*SAPq39r1EOYW^8&oO(vY2zr`A*z3TZQ|Wt&2zAL^4_boc|eaeBBp!0(Bef z1qGx0R}~=&^j7@W=^`bd@yP$RDcZb$zWx}L8jA`CwKw#iF*r>!{kQOAv;NpkgHJvA zw6ti^40W54({aQ*z1LwDztz|VFaC*2`-tKc1z5(^;?~9{;l_>YuLihpkN6}0 z-NaabAz8)LW4oPXPHihqrsg5bQ_MYATzGrqK59>VjLjJ?t<>I4X!t8utdJf(dU#qK z`~Gj+zD*u`>@ft2J{j-*r;u>3W5O5{{wXB8s)Oglrmtg3Wc#-52$X#ff;x|r*Is)~ zMvoq8huK#wn(^PUW2XjgKZ4-ZOfkXNdkQ*w`1&xOws+%m{Ff|QA}_u4k_KBga4pmu z;m@5jM-Djw!4s1HHf`D<&%$!#90If-J#rZ4k=$KzbNXAhbQyL*{34%!`Gp3sH+^%) zgij{|{wbhka5JWx!DoDz_ulI%zac0%)9sEw`~T?|pBXyTnquxhUnI6;o<4y&e(I#b zF9Sq>`Q?`~VZt~S)=oi80>+py>E`%nnOo8S6}ajeF=B-9#l{l*3Gj%QUwT=#z}1qa zL~sAEBfiSy$&=-scix6V;Hn0YreehKSo?%+A&$s0_dI!QTO5!7DJU9x_IwZR-eLd| z2G8OF{kozMM-hyj$qRqt#EJ4DcGc|Lzh8OrJ}5_Aq_eP#I(u?mjw2vB z>+|B@-ev6;{(jWi;gLs#N@Q>H!ngT$EdB}pIgj=sIJ$+;x=p>l)*U}~TxQIig?Jr^ z8DoLF6=X zojP?wjFslv#_?`cPp2>tzY4e)i$_V7mj&*$u2Z~wJE{&?=Xc@BKNjUS7Pzl(qS zydS>Bgm3;a=T_tId&->o@lQPXsL6L~^ZfL0F>x#YPfQwX4pWlP@uQ3wM0gJz4R!L@ zJO9;{1|McpU@XZedEvu8E9hv{?bprAKxLbeNYWBs4e{cV}{xAr6h77O>L89v^rs(?58ltX$`3nSTdD4Ha zI(4Ns0?q3mVtmxDT?>Kd>%iinj-E%wzj;u9@e}#|FShW#^N;Wc3>csmA*^-sq)FOq zZ~x0;^RuD@_v@gob>Lddv>9!6&udG~I(1MY^7u#Be=q#ltpEJ(YTC3hVo*M)-=a^! zf@SgI#ZZJakM#X_JN=IheBveQ_%kmq@b8BY=3tW&`+w>Ek2Ar}aPO(j^YNdX@qddx zGq3EkCRg*2W#=4o?Ri|TzpxAe%TU;?9_M;6=e8-zRFD;^oVn>mgOd~W;O&@?$*?!< z2_*NYB}$akAmc0(j2)>I!WJ!B3_IcB%N>b9nKmLIbJ}!iB}=BP zxIU@7V2Kg#U|qj)1A;c?!g)biyKXIFYNXa3;%CCcoRYIjwa z6UR@;#*LdKOV&(MyjXG9GUfz?b(=PC0-iEUnKHg$;Ttw=z?fc_@Dt%uyhL%y0T|o2 z?@*tO43Nc`Bx}&;0PKXP(xoigkuhT?m;U&Iv<`kA&z(CfB}Euw=@dSx$t9%bRb!Db27D7Dbs$7caq5sf>id zw`L~bAWz=hI_7+C8Rw6m1_Ht{#5&2HTu)0iDaV6j0uuUM&~26kS(dZk(xlml1 zc0!Q6U-BTn&siv!^5o7V$4?vw-KGH@mJtR$U$Sfo=0-Nub(fSbT}Ei}#DLe;t5?@p zFT}-q)K8q3M4XWD@RPV+FM0Fjk!;zrD-6OWEtV@?R*r(!w{72|LE%c3E(JQ0bC zsIOv$iu!UhckVn5wp|=F7E~}ud&}MIE0=SArUjf*k|H_c0VPW=6)IMQqG%$<{{r|2 zZ~(q^;oNyy1^jW|R;XA3u&>Dq@T$P{fkNK36@gvTgHIhfdPK*X@lQ5x+^9j}tJkQm z{zuoYT_ZFPFI>2=!as{Sv}(m_p}W1k*j1l3OBSWMa|nta4xG_nE@P$a-+w?Ar~v`A zPBHw8_|FO)778jTo6+BE*G)ie&LgHHu-e(WZL88~p+W_S+iwR`lJ6@()xKQ&$0vW0fFg-3K{mlOP5ELw_Uq(4fR!)ipX2D zdNt-v5ycnR4l1^YBl17;iYo~8y>jJBjpf0lrg=6)&E9Vl0!96N)+f+aBK zq`yN@LXlQ!#giL+$+sE__*JOlH2l~J%;%!uA!&sQ0>(+nlP9m@aR=so9`J$!z#(av z>tqCIt8$gf80!?^V{7Hyne&(*6;z3M8cNV}uw+}eVZAD98aHaBaY@RRD=+k~nIJc=laDE{_S+^Rz z8rDLk%aoK=SZi|T$|>Z3YuBvRF`@MnEu7M)577Bl4(`RyoIQj313-7BBt6zgww;OV|E0?pbuD81t5&T73!uLw{CK$XA+A4Mv&l)w zV;P*W2xx)4iyu6vz@v8V+^+NIJm|Y>m8yzc@;|yC&w{nGAn>KU2;;=}k4Wz|?7Ddg z!3h&z=1>s%k;<8yymO1>EfJajCC2-Yc8@K_iG?4jUy0CvKYV9ENNIljC;0sH(I4jp z2Y`QOe$c{0dBinjXhG+iECIzhqd8{glaJ4FzT4+{1J{~m-Xm-Y7U3M2YqjzRC)b-V)KT|B>od zHlROFh4KIX5gq*Z-g{qqVj~+AFyJ_YP3t|}k2%Gz|AzlX z#{WOze{~_${)MiTCQX*lAR{5OCBL?qoj-5BHaQ6W$OKtVO#Jsc2@ecj52S;5roT8} zf7oU=J4*kF2;#R>r$?}9=^I#9443vDI>dVYX#v?qKLjNI1tbp#D8r0pDH|1dM{m{v3?Ky=FRWb0P250QS$j`pDA$mKfc%BQzuWu z9oDzfzu$MNn4?FUF94qk7{aGzynp@ozyHLvq6bk{5-rks8VIFFco^Q%?aQ$2E+qrN=ghB5(zJr;A5))~L>!%u}n zc(hKY|M@3&uD=f<<{}kzb2oHV1Y;ib;~@BoFDrL9Xdug$EmNUl|8M%q;Gc#;uzXq; zL8Ci&?tFO@!JJmCSPr3ED(U=4N8OoP3c^nEata!If8JY~-qS>S_4-)uYt>4HrQF#{ zq1Zg^l%+7JZ{M#~;nBPImr@9>Kf+I(fYAArx{o;z!DyP)Y2|xpy=x$l?+OS>^W<%z z!F(@YzNQ*xx=kAX%LoY$$|qB&P1Rsv_qV>^P$?!5g}L*v@Gn?6Umkei0UgcN>(_!(}9PzNdoFYI?t^^1=Hb zNHHixFoe?oyM8Jp?9->Wj2}NynLGQ>Hq#AI!9u}M+{~2@?c2+kF=JF4|G(#+l`p^e zQu<)GJ{2I>5Hxq=rcLrBwB?4}lE~~ivsLKH-P0>qtWuXM6X6%KG=yedyL5%{f4B7S z|E&t^zxlSGbnDs~eMPJn2#>$&(^p={*vy$VN8M6IKpn zV$?`!-mEFy$%H}Svt8Z?O-;ZK_GV4*k!jPXLP#5+N;B?|W^E}`rGV1rID~-dpoBVt ze)Lnk&zw0^+O}=0FgYhuLLrenSqccRvqC9!P6KW=Zq!g(wrD9o{q(a63b|AM>1Upm z;lqbQfLd67|NVFQ=);dxS)=b1cp;#G^p#g%mWB-*Dh+l5tx~vr0m>Z;^y}3NQR^Wt z%;n3MhuP^O6?UFFc~aIv;93>Rgjs*ikV_b+HS1QZ^5EjdE9&m%lTSX8VLuI*9|sMT zr&2!y<;6M(yR*oUp+lv|6R^NShsi5!GT1~0ZhYLUmkMQnf#nA+m!5zAd8t;dn$q0| zAAYEFG#rW}T24^m!D`4!{(`Hz5hF%J2wF%^L*ck*?|vw>a-ltF-3-l3KWaPJhV^LQci;Dy;lqZ@H(!4ZoE4OjBS*<|&pjtY(Pq9J_v_zZo_y*_c^6~N zIWukAG#P|CzW?zD38`C0R)TJ>;Z5fE-zTDP*P)m=D#7?jpi8iEfZ)Fi=m~(XV0Ej+Ru<70K#63aoBEEJ`5c?L=}(UegB;d8!`-ZkVXZ=-}d_k z3WlB-PYCAEpVi>*<6#9*uU}ZHXMQAo$VSpU?2qHS!CSU@$wLM?|=Dq zU&)m-hg5*VwG;&C3^M*z->;;8gS+LMZ@z-n({u9Shaae7=-dUk6omB^6~Iei-Sp!R zKPVovK>?BliZ3cGeg=J0Fp<7|={$VNZiOXWGvERW1@wExvK7*-`8}%S$pD|Yqz$^(YTK@j zeDTHS^3uyM>%1e+o`JbX6KnEI)a=lj$f%K{WznKVQXIi@lpxj!c3Y)$A_Gn+$tX(Z-N|jM%5QB+d zf^tUZ(Xe3}P_KOXic+X>5h#qV$#V2z0>-H$Ov2xI=WThUQ%9KtT6pu#w`C42+^BS6 zKNc-s0BeaWy2C3C6oS+LoGztGlmZ|APFl9Q4>UO#>&*Y4)IN?)0Lf+8u%S@w1Zeq= z9Xo=b?$Gh1LJ`{&-2U^8{V)_>gtZLH(^p=8NtL0&!G&O9*8xKsmH?$5tpu4@k?FEp>bvHEAr^GyXq^zOlX_{Sgj((Bh>e^c2tp)%dTf5~*5ZYBu- zVH@oGw~J?O+TUA$xH#$J0{I&nZYbLH(!2gV{P4s0cKZh30Soo}?=5T`^R7S4Cl2}j z;q8CTny~zU!fF2e`SQ{$ugVvne-df^I~8$C{`l<=>Cmx#4Dg+{xFqDmKl`tV_XyNU zJnfF4>9wIqXy3kLr1dxAbhs0{&uPDTxnDCKGM)H49nhO9H#|EG!gO7*i!*{(N%JJ@-ht^5rBacF=c%fT(7T8gR+67=o$7((a)LiHFfm*Q%LOVExKo?B;ST$y*qKgYAN?yc8Bhi;NTz#9&=)cdN$p?{N-1D@O%(T z8X;g1cI;-&ngu&-ugL6KGo?X;hAIdgITEh0qTrtbumf;m(yk2!?IPoT9glY9gP`|* zZNqp7o({up(|z~dC-1-izNCP#K0DyDZ`?^ucU+UFPLceuprEft+eg+_2PTR;Ah6xJ zbGQ1JXCC1Pqp!8=)RH!Uxf6o6rEp=A8oPJvqTi=aMyOyoTehsys#QxBU~#Aa2(v|HMOa;wkb9aom6;H{Ux9Li%Cx)Do~YxW6aMK)Uwlz==gFmRY(D(p168n5keLEP&H5oBY86BrafdeBPk}gh zb|;7B24exx>LL_^`Qt~AL#Y=|V({?|JKdf6&leP0BvFY_5JJw(7`qM~J3tXx28yL! zG68g1uYNt9JLwT59jMd97$jS_ZH91jmhNg#nIbu?3N9-C=w6Qfr=`dXFFp?iP$M;) z<+Gd?#fle$kRpl37gz?%tHzBRtMV#+KzdaaF;>EZ54DwSXba~+H@L-Pyq!xB-e*J` z-USY6L6E;dUd+v`X#XJy{r}Phhl1ALy?aZx>{->qga8@KU>a~;0G1k5j^zc;d@pzg zAslyF_b|2z6*GK+di9l8p!h2VW!iLDQRR^PF<0}zij0;ntuRh)+TIVr{yw!*`4TWQ zWXJ>@oL7NAEsGv(_mC<*i2pRHQyXg(2%Kp>K`RAXsBqUkgUyA7g(>~f^$>UO7b{*A z*t!m(_B`;VAZgO1iR8|kN9RBJV&~4C6`4t~deEw28~E74gT|#EV?*5kKx^>MTHs^5 zz*AbQ^#iTxvVeCo4iWpGB4tX@T_IJ9(1L~I!ktMGaMP6`Q=nvp^++&a|2p$0C4{5ta?-U>y{g zK7(8aopAi92x0%}KbZa3+PNFafkV2al>?Qlr=c9Af|B(A#+$DJ{&gs?4(t5qhafE# zXsPsQm(JifZ(#0Bfx_$$rK4(4C>8^4(W0$3J|yVsD<>37oZD0oaQ@AlHB0$bjhfXJ zK9xPRz@bH29^gMW)`MrBenu@~&R;ky9WhtvUWy8?5>R}sqLL0YR0))T?aV39z3{v|3={9! zvu4Yqk3K51X3vnBGyjBQwgqU=7qf+HLrL(Z=JzyL{!99!CDfy!neY1bhjL<|bnEiC z^R_{3oALC%f8@%8br19JFyEYk`X zE~@!KpqT$-L(IgFJF;G3m2*dQeN=pXZuDYxwU= zWF}5B`JTJ>o*rZ7f41+o9`7-s;U{+g{qTMCL^|+~e`I`c@_A+$5#&ynxR`&I{#dK! ze_AgW873TbOe|Qg5oiPN^3%;ILbRl|%}jk3HuGQ0=_SstVd2!I(3~?AUD$Y3)37Nt349clN&*ltMaap7n>; z6rX(xO9`Z2z!SZlJl_3r=?*yzJsvb^+C-js;tBmeirW7H1HM-ULw{I4@cZKxe6z9+ z!Y_&IC%QiY1;r$tH#$!_#3UMge|tt-VMoP3GnGm#k>O(UAHQP*-Ot*dlBJ4c^V-V_ z@6}gcR`?9$PNf6KJ-YRGG`_7VXY2LR$M_B_iT1v%f5aKT%h-N?uf^5+&kcC@-rGWz zMoX}P`spXB2x0&I>%R`VQh}uedn9}CzyENqMWz3!^Dpt=f6ORqJH4x;T^r3L^xZK- zd{h2!YQ!{2g!R{K0aeChy2Qq4n;j;^ozFJ;&d8{I<=tm&LEbZ!xH*w7okw-Jol3p+ znX;x(T*K#XTg!NpKsZRa+!22o@dfBwhAvFlI|`!GK*&pkjF2=Pu%-#|YYK zwLu{fg@qKBjT${#cbS_GCs6@>g9Z(-<9i^K4sGDFsu*@Va1U7$6;k%;^A&`KZFCns z1NY(Y)Xc+CaQUYI+A9oAn+NJt2Fe`Lh{Q!0!jeBzmjJD*8H*#Kh=EK*o6UA(Y7ceImE z804D$uZMcLd-?wR+o%9E3<63Dk*{Fq{7wkRD_5?n)))-#9vBb^iQy=%7y zxK08(r7(TN#*M1rBHbQ?;PVJ9UJOO?Nrp3n`xyxIshA524pRC$eB>y0QGn)wJGEk9 zqmX!@U_Mov)Y=|CvllL{C~zdrktV2wI|IQruX&^;#s&0|u{mngtf>NHS~f6U!#Fdj z`7Vs1CR9KgqeTgKc~F3Q;oJp7Gl0!LQb9!cbX&!kAG8oMtrKfVVs|URTiJgoEerl z=!(gyaK4hpiGK>Pd7c@x@Z*m^ zVu$`1Skl1l62_imM^{ve3#}dV5I8;wD-F84y9x^*K18_VBlem7Plhq!ITZ(O+qRQ#pz-$DC8W4y z2bfS7`%hTxKkwl`o-;4Y;ov#!=XT67y0$wB{=@32=p=kT$M%y~GL92@1@X(VqZP;z zC?Tn!xQg<}Ft*8{G~@WNFK2P2#md433ss47<_G<+^pzU;H#k!w(>~ZN0 zURoKhz!b;GV^ThfPLmduUm}lvX1nskFEZ(G@ZiDH07}q6D2rFag;>Etg|toNPl_X~ z5$uxI0V}J^S1#cy2|ft->ikX)ywD9Bl_`w#Lj?jFfikB5igcWxhNT~sO_eKEl2UL- zHfYcwSOqmO<@6VP1Gu95v*yj4D{YbnSYK0E>@e;L`=2#i7A?pA(B-2Uf2RT*rcL`3 z@sSQmNx0YIrULSRCZbLgPB@O-Wknvqcs<^GJonALBpBajG7eTG!G)kyfl|uiKQZ1v zwEYIPXfsZnhTHX&_chDnp9JqD-x<5$R1(cS?Vzc|KI+uzT>9ZFBK-LG(_ggnk8=m} zPnTq_O}Nx^XU_53&ObXR^%12i=UL2K@^zEU`xom_Uds1B)ZnE*V@1mQridTDAOC*( z!(2${`G*OpbNWBL{+26OK`rGCztXn}?~lU;9oFVKSZ-$ukpJ16=6(L%_ue%bPJVpz zKUywqhMa>NHs@o*>W95Pf@>!#5(JjjZ^3GMpz>QbDklERcj_Qm6i_yx?L}EOKR)2>8JmKxkoheXglF^`@{%y*^ZPHl z_2(&ALrkAGU4{1io}t2laXRV7n&0QsA;(}_qV}I{;YMO^qT+M?p`@XAfd$xn#5N8b zI6#Ag6IFI&0FYra+5B<^QaUeJRUh`pT_q9 z<8V+W$M^rJ>o2{%Or7FuFJmL}S;k8#U$(r;Yc#kzD+BRHg&(#5C}o?kM~AGq|M!5aREFigx zjq*RW@OyQ7Ci%jakR_2b3ftLcg~0c;*TP$I@7cyJ-``MI#^{$4z0ZP$H0*s#22eth z4?72~NTiDtsQ81PuMvkZi;MzBd{mKB%TPbiyI|$_ySMkL=}HF^7;?=CFUYnD>|O1Y zt#;=EBt|!HXjGsN8_2KwiQ=R>!SLzh;tj%tovEisrIs<;j28!2t}Go*qz$tCScSBk zdlp0Sm#5CD?iZ_So85tg1JVEny(et6#(A+FKRCgOnH<4Tit$3yuJi#Xme}|16T|k) z8-E!S?J=rxvn%@k1Z`rucS^0u_bO#nXn)4)4+;NaL+9UOOTwhNCT1e{#O|&nX@}DB z=lPwZUX_WLH}?fx8|sxs#CsB^$8+AqV(b!)6RBbG?-z{LH!!$j$De$l@&d%`e0hlr zaZpwY8&RF;J!aZeHjgAIQ|Y=98mlFyV*uYUOON)%Q!1Z2C!(Py2EEJikO?LNV4Ze^ zH|n^WJ_+JY@xDGht0OyVBE?0#)}^U(vE6_&j7R%6RSoFLtHi}0O0S{{udA&BzYk?=9(L6Y&8kTQ ze0f2FP0!flA40@#a7SG8N9as*@ZzB;X!X|v%Kjq~f8sX5OFJkL0Ok=*(`Yf{wsqhP zq#Ika?(6#R$_dgDchkSuy<}8zyOGu`DiqOeOaDF*GzFamx`0sQhDND!Wut2v zvoS4yfr&wff91%UDYljTCIi)>!pv&pK9$f^)l*w%E`rT;H;jrv!KEY1%v|xFRLG`b zp2?h_To9n+p8AC$ceYe>e+1)pe!)0=&$O^mAT3>OG+Ef!#&6v87}pEU;|d%F(4P3@ ziFk~>U+IjATm*e(KOzfCpz0b*Ql@wQ5S~jL%ltJQl{xK~^%Q35iL9+E)KLyq0grL7 zLNVa1@BI82u(?{fqor?pMiohJjpR9Pc?H0Pzr(QY6Vt8F$SHu>pg-7rjku(L9}P zHG;E@z+TSFd7m+<#t*BB{bd(!s);R7ud}t$DL! zyv_5aFbRj@yTgXSlM2D9z#u8{bQVKyX)JS&IfNJ%q*D>)XWpAm;Eae*#h~POVRnxf zuD^|Lon1HJn1?$$f9b zJx3JzS1fev?WKp{<*&I6dyv?)aSR__geJ7;6IOY&b^{wSIw39&MN68~8|{mm##A=Z z?9qG`n)*qEYvOZXrD!(A!o&Keh*R2)#cj6;uT~)f4U~kYwvEfR1HQNw0NIgs6QR^A z;k=;UZR0*ll?llmSXjmb9@c#0O0pCZCjWIWBKn_tR5^#s9@2y1K_M#TeDkYFjTB7B z=5!A--cTPli%wzC#PLIRP{GVv1MvF${R^wgZaB{K6R+hv0mR+9*s#kY(-Q2vfAJ)$ z(bm<#7D2I|%6Ug)>uUG~YruotMVv8X;vhw;fiFU8}*|8Goq!_wnB!1MyB# zKm5Zt8T&*g)^f!7UfZTPieB+Q&00=7@MWQr(CI!QSuHu(WE#&AX6fXEVz8Pj@AY3V zlqUoNBgL(-Tj&&(2MpO~p3D$24)Hhqho1bm?rCT=c?|H3-7-LLw#rk|jPtG!*P>coX4^V_$^&Nee()AW%v+s;YEZ)E|GZb5s+6cIh=VN}H*C(HTzH+$^O|M0sK^3|mf z^UKXf(DSDnZkh#xiX_7NC!S^2sQd#ZZ*sT#Mmg1%duK%6@V7)w;zFctSBh$Y?oQ{I zy%)cd4o{5Q-HY7$rU zE_Q;a|KDu;at}WR7qh3}(Zr*O-AwL)kWDJc#nk(1*B#AsUSk!4_RUt6%1M2opU!RX z%I#`VDQ`os!GU$|;+|-xawl%nWB4KBy)l(V>4ps*20BKu_>Ovqd!s%z)IF z-CzfWL(;$(`1{q@L~OkTdmdYKW7840kWEd^%@EmBr_L~f?UWJxtK9SI1is&+kyAg0 zHhdNwQW7FT&447%`?(fZ^aGd^6ES>T)3ZT^9N3W}k3rom$GZm=O5tR3lF_Y3#++3w zbqW>qTUh|0iiGNL-tf8{D-rgLEe!A}6$wV$vPen2s}Jzo;!uOX*Le@^e>;U zcz}WVr6fQ)K!;jY1zyMEoVv@f-K=y3Xrk4GYrAj+?-7^E=)p)p$dNzbt3S57`8fFr z{2?kv%&sW0<55hb0JTeV4kLXJdBr~S>WFp=$zAjax*Iy1WH1i0#oHw0+X zmPN3wn&5^DJviz2c~cM(P&S#E?QPJ;{T9uNa_`aH%W_PyL6$(K4d25*DHnsb!Tz(x zQhqFk9kFn$wo||V5hWNSsnUlm&Y)5JaiJVq!uZ_z9UCicwM3L4lHlJpZ`9Z zoR-ypkl-i%7C&Sxh?iF1ZdMrYG(j6b2YA7mYEos72siSZO6gpI>tuLx=1+IYfZhHP zhO^&e8j6zjX3|damfKFZrrEE4)^2e0NO++ZT5Wa)ODOH?rw){uEVTS_!MzhWkdjwM zMRxOQ@W2t@15vwW4bHonVn2D{q^Yr7y8H1VEprrQJo|0mhn*aGpF2eDuNmJY`3!$p zk1);;S;ObEe{X{%0%5mMH?<2(($OLRxj<;)efhV@gJiQ!#-pd zl)nm79uvUxP9rm-kZ4pISISaOR*vRuKNRaNx8&m9Ngl=Ng3+W!9jjL2s# z`qGg1H<{=7R;vBG&TzTXlM<3Y?|NFIs?i9Uz@)ozw&h!rBt#;4)PlS-J4h=Bf4?0o zemI-OyhAyA_}`&r2Kuca)VB{8$j1{2)7^D9sRWY~@s<>Ir|lvwPae29*T&*<_VMp2 z9>k|T45qzetka+w8}zE)KFLp_3ycQ!ggi^$z-D>y5S~9^c1%#EfA6Kb$9Q6Bo0f5F zl`(y0H|Lg~f#^rpDzFC{CehaIb1!|fjX)n#{sx|zferKWfp7Qqz@6~hPRQHllZs=B zKlsP!p<38`Vj+u)!FfFEx4C8sgPWP1YWJl2+YaB;`Dr5Oa;M3o{AZWITt79}sZV8P zWLu;{9yhj5k6lL^FYtv=xTqrGVfH_Y7S_d@as?AJ@hQ9ucN`7M?pITV%#ZAcnCWXB z%^vi&`s9)Zq!wqMV_18FGCqa7w{keK+dft>{W4h;no38F0bT`J7MU>mkjn6t!ul*F zl(K7zzSa}KO;7Gm&lwSe7;jv|W4gR3Bz83C{%vayZEkOsa7&=cGvJpR< zYwIukcrbEA0`l)$2VTAZM{AHfwco52mLY)g<`UTe2o=WiMqV@m;q#H1X4?v<%Ha<* zCc&@%*nDhsj4zm5p`qVu;2P=zj*8cf5xRhTs+bMgnF6L1nnsV%GtLc(gR?rHnb)k2h+)AdGB0}zR2$xc zLlFw&NZ9G4ZDGB6$a-1Xk=-QcAgCAnF6bMU;*N;Yq!6?yZz(E++4&zP1S$$l`~kEE z%ZyZApoBH(Vgo`Od!z8!=JTMJKfzg1!5$hMU)>h_Q+!_d%&*EQ&Z!+0TuGHJ1pPvY z$p-|~4^pV+Y;0=veKJ2Ku=67Zysz(>f<`q2LR^ftyt5xUS7w3Z2Ug^IX*2T(pF2uw z%|vk)!c?+Jks5v2R4$Pzxsc=fwit+>TY^;rlX(n%2Z1>V3=B+OKef$Sofqw5MJ<<@ z>dZ%9yG$g4VQsb*a6R|Mx!f_LdFxr$R<-apqxl{N|Be7h%}v*%O`~8^$mNElR@Hm9 zR(D6S!*bqJ&xrx+2(`7{&+QEMyLP4)bHY@XdR0v6f0`;(3X(L#8m1*x1=N!~4{p@7 z8f+;*KNfku8G!urZb&}+zpHuL@<){L^;|RPgPW^*(43h}F!BSj;wXzGa=gQ*t{D{B zsSLdz&~{1*Y#BJOI#?J5ndf`um8M21eG2?QO-=n+h{j;@J4awjwgavxcm`FMJNcSt z=Dk=941oSed9Oxo>&`b`F8`-n98k7wnrxSoMZPm~ZEXFzt9Y6IK)f-yKJ}O0n(Uj) zdiB!-xpTRlE#8^~RB-J6Xm>bU z&aKgBJuCLi4!w@N9_{@Zmk)_Tft*gx`bgvq-GbfD+`@4PDbURSBZ zeD6OqP%t>52@zqhpZe3A$YXu)Z6B14{V`w!3NfvingN;}h+6TG`tH8x7JA$6hR|5k z(JMCP?z>*Bi<@**JT(=ZHW40wlBHd5pgkVYwgm8nx2^J549r!&mpNig=M+84vrW_x z{34d=9R}=tD>o?P#nxRJ_je%s`whk7YPmVv-KQxAU^Qhn)6>A4^a|{}(bM_1vyHVK z8~uzdP4QC9F{IR5Algys0|CxdUM^$J{eKq#+kC)C-p=2Q6DYi{y|YN><_@L<-_6v? zub;g#&`gf(bY=7&-QA-3r|h1@rz{X4?YsA)-WZC_7bJv1FmJ4r+NaWqO`zN+uS zyNu|6ErswGz{lWfWCafWWD@+s8_^u7vJ}eS#vrKIC=}*}QWN6SLd^NyVsPL+qEF@$ z1A!WXw$3pAWb7_X>7av19hbXEn2KU0CV8WI?|O$afL2|K^|*+E{(GX0quW@N#Ii%r zPfmlqEZRBIAR~7F5Y$O|7KA_L3SaOxl4K8o8p_)a-e*D+h^(2|=Y;GU)K(_KUCSMM znLcZiqFQk~d5tXWDuUb~SS;`6u9@pJrw(gDQd%u2!t1`)eNd)Z?DTvWUUJT@3Ea^_`gKZVI3Y%$Z=g6bhERj+ zhOd`J?Vhg18wbRkaP6(ox&9Sb`syd9T*e69LP_emP-pqwKOMg!|Wa9l=0kE;~x= zc-~AqI^Hv-@|~$0*w~SUDjkWT=%tdXVDc$Uf1T+fE;l(ps6k5K&S;67Rn^CiM{$;q`>}|@Az6LILTJ_oF1EI zcIG%i#mgAo-Tpl{{8th_^$-?dvkJQA=p@4(+ALEh`f)v8IL3Qdkm#l4iH_%{`;GTO zN}vm;XJ8-ohfnCU!3_zebbxQ8cvciaI84Qp;k9>n8?ne=M1O*vMtm0-zB<=^4BIK{ zv+nGAYxNji1sVtUE&&6WNI|reyiGV;$~QuHAiwK7ez(g5>F!E`(Rc`MCo~X5%mf8> z;ycKuBjq+c)cXREmA8;U-{oWiBQ+o{Q*3by?bfRWX%-rr7N?ndBY%b-B+q-TwgxpD z_7}?rScFnO<2rp5)4%&KK8dA%vtSIK7R~L>le4$-{@ilWJ+XK6-{onXze2kqSMSgz zH~OMECbh!&az}!1NexqiT9J{FGr3uC6F4_CpDFWh??q$S>lpZ+?t#!38F?y(n;BSM zv5Os^=zs{!&N*yxgF6O=SSE@+46VG<-PW9V1zrX}Z+-Qw?~2VN&o}|~b`w?NF6TL4 zLvQZ)+#}<-%I&pFN(2jKJ?PB^2v7ehvHCCFeHu_a`-9bHBhLnUc!k0&_T%r6rUTIe zXMS5wpwNbDZ(2b~4zl30$haNpEhkdq7KM$Q%b@G6P75Y0ux&Jr@Jv$F^j6ejRh8je z8(YZRvdbEY=uNvqm;^0gqrF-{;Ye}5%Mu-9n(UqU-0b#%R18%~*HO!@_QNO$4o2=- z4jOY$&?@_;R~baLYAM0`n^mPoXZ$P^-x-MP=rSBz6x51c6smGK^Q&~if*(PXQ;av2l6?;9g4AlI z>cglGOF1x+VB0W7*qVt;3rM4ist4PQZ2RSlr;*%b-G*^|YV+{Fp#a9lf_c41hqMU^ zAnHTuy3x1QYQ-HL!oLrT1V@na-K>^t^nN*Qi^QSh!F_36^Z5RSHCA6m6Qnkp|KNh} z!n)S_H2n3K#}Lx?R;wZ}?TdLk?5k#~6_54}>yX<14RPB{cSHdm-lg8H^!H*@mTSd6 z%w(A%Zlki_{hiv${t0w~aIOmO$1GN;X&~0gL>zS%4Oqa^OuszVc|1rmM1e+?YSn}Y z9e!p%Y;OPDkR%4~IrYkI3Hg{!0RvQ>gnqwmH`|rRO5VIU8g>9yM2w@~Z>Kv=vzRzs zZuj9L^!)W5{om6stL3=#1|M3_ZR?Xt`@1;839*KXKo7M6W4(qJ~YBIy6erN%Bbp~-8$?q7<>0bn(pk^a55g|f+Vt&-A z62A`|XnXe)c|0NA*!vxP;4wAm$^*u6E@e0<>}_aJrB@+j0FT*hS1y#7>fV~U{@c(i zXTk*HL1A&UGg++Uf<%r zOi5zc^aFi)pd_~6$gwHOx?)Ha`jOHl@Ypi`CNyDgn}l@wmANA9Hf#hpruTO_9lhI zb^UbNs{UBVSS5Q4|Clq3AaeZv)Z%^00LE1hnRwrfTmqdRTB=G`QQt}z@r{Z?lveeH zkmQ(dA1|AtnnCM_OBJr0gnT}h7=EWDv=G~gqbV3GoDc~l@=4RtCK?5@^hL)KWklQM zTxiN0mqtE~Yesi~LkG^}i@5areW#c$=ca_clWLx$N0t010<<;sh9xn#jK@_ZBl)FP#XghgaJCl40$p*lky<2w>}|yrLNz@T zNV(-xxAgPhOiQQpFSXMq^(?L~_=)N6bWs^Gr!&#FwM}pHNpGEb#7teV&t<;N0OMur zHrg!_z~As%lhnCD33i|hp;t;~OkC`qf9sejQX!N*P8MR&^X0nK!d2Nj5pa>0&x4vR zGPNpY?s?)>S_W4niuT)%Vq-N*CTUP@?RU#=KX znXDkWuY`+JDnyBBC}^MSvFYX=uqJ64qIY!;`MS#9pU>s1siln?RcJS(*%k*S)ojexyy44u;*|peD{5gpnRoY-`V2wAN=+U4i|PSTMOJ0 zymC6dsKE6NX)cDG(b(mjd`QU1V{2M=C7e9zIbXqY*-?>qmJ|Hx%Dv9c(~3~Fs6&wT zZT-_6aBAW8NJs#3uATM=JPL!51rm?)83ARH8h4?v+MfT8tYMizizVLA91h1S&n<*) z_bZ}dahBD&w0_E$=s~;_57?GTmFI2;*&;c7-#_@Rf7nFu^h@R7-SWUCV$a48LZ2J+ z_&%`fx$8#sbYt2pyYyY_@CpUP0(4`FO-_>x^JPv1zoA;}A1RSycWc+K-N8nqa^cab z7X9F7SC!vZ9LSW2ab(;{iHX?Z^%9d_+XVU%s%?{oWiTUr@^r4CIDI}w|C03|SXklE zNS79$S^ev4BK*LiyDj2p^c4099j0^|1tNv$0|qUA^%@)YOw{ub9peB32|c6_J-&J{ z&>iC;PEM>ui30X2S7&*n7X}+n;Bs91`cXMIEiRi!9+p{lC*okXKPyrW zd=y?gRC^cxj(Icca3Y^_#V|~pSQLX2Glx7eQc#E~67xd}%xkRh;tt>DB{vdBQ658a zOG2C>zU-vSY6=SC_c%+@gu~3@vSm_8XG-?Kp1!xNP%o8NFO9~L&}dH;l5(9=Nai6} z>(-b@E?X#A(ewlhSiQe-!%h>!8aHt2VLnA%sK$-i&;MfWXlfrST6Sq=t!4KB`{Z!nd-n5`?^A2%7Y#j)VCuFFwp)i4C9 z2B<9;z5z|V6tmn`!Y`AQ3hLLm?ab3XG4z6OVS{h(Z$5^}u%&aXrFLDDS~A!o zo}AzEcpBFHHF{Lg-D3=7hS^F|U`9zqEG3Jz2iy6AK*UO&-n@Zh<(uIcqT*?Sf?@ah z#{Ewm7v=doA0mR6H1-Z3aE76)9pC)2;5#|Jcx0Q!oJ+ zoA?>B{4{8W)ZeI4k{FxG!4UgNCgIOXQt^_n$$i??e4T?JiKuDU2CVxkdIA>praJKj z$mNUF&O-t8MIX-Bv}oZ_bL`e}=wCQew8PN}QF0~qkS)z%?%~6JPb~3_9zg$r&Hw_< zyO6QgRNif00{{UajV|9;|2&V&=NBWQk>4S=?(W6UF%8=4LnY=5PfF3iE`4YGFkI%3 z9}%=gI!glII(N6|GixMrBjd@F)UkbGoOlW8cyelo^`_+NmHMiX^2fKQevXgLjR&cu-hfg_o4Sb}$o?S4Cy6o99dQ!TPzuX5uFuPVCMJO@`Vx|V) zq#=k-B`*yY{2nu!43_Fs;6s{(IR>j{HfsbcPzu@JoV~A!Hl0FVw zY(oNEk=U(S_!DBCnmLKk#HUJEWKD?|gXZtZ(?7>0u|(e^ zAM(isdkbhfx;zOJC?4>_Ztlu0t9la~XdD+-a#buu&h48L|MZ5{6x?7#jKHa#*d7G^ z%#;BOCQ^8KeeL(X#T`CG8+lw>G13q^m`GQN{~*Yi?|=8sTMQYE@e2cVv)R0y3o!t8 z7WxKT&Fa1D5x1rxPG7TO-sH&Ve2tMH(HxY(4>s%4cg1xJeb*IgpWGoDyYu^FbTk?x z?>Pe2^5ucCkHr8f`k_B$FlcwVkUb!!I+MdPak-W5XSNhW@gM<^`l5SEXc%tj6aBp3cEb`5_RCGa17QQ`;~aUVmKvR<~= zvv~eR(5GXE@tHD_T6I3ycsi4lNPf(}PsOeVIaq+zW|-4yT8Ieh-@QjVAySSyEimZI zL|}8d_V}fspIg@;U7qTW5>E7)@VO<)(iTWoU3G8|80^d!52=z=XDs2)u6_po$?iiGIcRi07c+-t1boH)4 z4f>D00O5l}47d4|=Yik$pxQT1;lR3HI-G5BR1A!XD5tsiB#kX2u3#twxZ=<#U1D`?vD}acxrmV=Gi&^Tfulr9yzog|$ z{=yQ)G6WH|;mZ?x7y0W;4UU6z0QPtg@;yV7M+^;`KpxRY7SepSbQRLzRz8r=78t_F zlE$Q0HJ-rq_t!XCy(F4mM~PxKhZ2iKlDL%WQfJlKs*57S>`H>ffiu`gm^5TM?wnUl zwSkQ9wDwBn`*Ovn_C~>U$I|gM)-TOeuqZSvIIg<8YJWjnhZ(#9!>dWR^bQaeTgQBOHCx}A$ z!W!$Tg2LWUK-@fyLbV6HZKk771^G+Mw$}@s*Li&~)u|AZs??hB9%>B=;Cf3x zmK0El&Gl($@@bUUAW0#zCumj}yy5%nB^EZU-NQppdo6v}Ur1;AjD*TrNN zh9aMstJfAW)qK3@hWtR%I{68E)qlw9|YX#3;sp4Y<=cQWWb-KqLHzWy(5@FPem1l+=Na)J)c}t z_Cl#isAG*VX|64GXG;*byu{}C0@gI;G8s$#UvA5-PGtw$Jq>$auLPTZH+O(vE?a+a zVC{E1oD&jc^n&)#s~RW-9b*?Oj3}Z{c2&=EP>lu_9oB;=QQ)5B76VqiXaDX=#Vp}Q zjDH$%TC8AC(c~^6#;49LjLxQui*S_6%%zZ|MBpyzP-hXIG9;-d2M zlRqS!1!bjnUn7fVjtp>WIqS#{*o91;(T>=t35V2SqAFi+`HE*n>3(mlc0@^ z;Dt-XY@``Y#7v8yV-zV7JO0CDQic}rO2*Yj7JcZy(^)#3X3D6z-WFJV=L86e8jU9S zg1}i~GRr;^7QI812UHH{ayK9GsCY`x51m;K!-bN2&!`@f{rzRNaBJM_*ex>UHylF{ zSsl8~jv3=#!w22=J$Cq{(7&cZR}z0Pm8-)j!7#PiiY0Utbf@>vcz55LVZ>o$)p%E; zRyalGlsjVK+^GwYgCtvT)Lrm?vid=#$BQ`nhvH7Fu|K38KS<(iK}95$+isarsi7yo z@e4n)O`Z1`u5R&mS^(-lanmVocHZR?%KoRL2I8Y9OA_8Z`OdUE^ySci9TH7^N&$D( z)cf8y5%}nqUEj@LQg#6-vYdF^B*0|bUl;A&FTA}_8)K%V{egM;q9@U$+nILA7V5QJ zccQPi@puUfR__^%UqgNFE`o$C&c5=CB&8)`fAdh1gyY06lhdv!>`M1jr&rB#FP@AS zI@HraW@~zBhuO{I-{iIhc-3_NK3C_EUx-Y-m3FxbPD|M>54dz}w`rRFj_sVgf^M*3 z%lX5)g_Y|;wRd6f*AXa%IM9m0JrN^pM|9;AgmbLMg$1Iv<%LC)W7g5MPkym-N|-cp zKu+>m7oB}VB+PA~Rmvv>+KLEW^Kj6u*)g`(qaBE&@n@c)MBZ=8UzxV`AB35)n*Qu| zGH+iqL;ijs%C)Dg)2)+RkZ#!fYwqe;3=>zZE#F=q7ve}@^;|!HmdY0~-xKsa15DqQ zbLZXmQh-O1pxU5(u2U=r<1uxblFvOpr-3}w|K-P=-cdg6>5=lA5SIEn)_(&FK3r|M zV|gkyw+8GxRo|+WGM{5duYWbjtR4K8T|E>+84<@vy^3_iC^1Vu0fjYqK~KvV?kTL+ z+EK|6{Ds3U(R!}AeG)~DpG z;(Ti`NxZVZM*pqSq5f_XJBXr0suBH-&vVKm=%z*>h4A(8?9eU>Rb84^g5BhGJl&D! zulsdJ3bO%WX9%a6%&zZ6*j6k>w^U(}O|jq@L)fnYpAXL_A10^EbYrc3Kk zwN2)JX=SwLF-fnNBj$`{t;Hr@KX}>(4n{sXUXV_I%KdQqxtyA>GdlMbW|kzpLg;lr z{TP<0Ktte}zB2~cSdRQ1qRAzA5EWzM zI}0pq5ekn2>0K}>MCh-3w>7>mSJ|3*?2xf*RpAe~w4)S?)-uWO7Z>?XeRsVrj`@D* za(JgTq@as%9>-{7O|roJ)?Fv~U@&M#YtVAh(WyH{N{<4z)YJtNU4s&T7_gW%cmu3Dp& z=cEt;UxLiyF>481+j#iuVUaX89*~nvr#znNhSR5DOQ6i@?WNuXz1fv?C6{n#e8(us zCnp~0xf7fw;3b}EfwyjRngGak#l~I`UJyqEFeKJ!b(Y04@y`F}OWrw&$B!zfOrBw(6<5%^I*LGx>(2%xtTEQk=U$^B3%5;mA#K$X?%{GrApOMJ{# z+*Xvyy>+iu41yp)m&I*Y7d>P?s@(T;--RHDFvMg0o>k~rl{U$YtwfOK5kH!Mo(Lgm zCO%i7iYhOS+^ryoOvpR23&tLLnVW*lZmI)xX_BU8zi64(mfR|dskj#o8(wK1HT04W zP%qb?PP;^l@vn=65HO1;?$+_oKl(8h3Zmj zvUpbApdEc>XiskoGK*ko{*&w)jFss5Hv&ClK|dLB``;6(=S#9xNQ3!LmP@{b4$pIh zJnZ1-L=rdu21eUz?REu1rmXt?75M1EL;-47HA5c5NcHnJ#S8JB^HS!5wFT7j#!*GpcfnyX2#?Wpun?(qCnSyJ0KVHd zV8~R>685zHE%;hRze`t{vsM+|M!ID8XFoxhw!8uHA+B}nXJ?i6AZST#B9gLju_iJ> zN?%^Bf4MR&#t zKcC8z_9iXE578|7sXbq+zzXKugfmJa1H;Ozz^WI^3;)isKn$t$wWf!|ZDiz6u$qD@ z1NLd|tApZ@T(rxz((3tR_Q}C^x55J?iaC#+Z}eWR_UlD+Es9wBgF0~3UX=%o} z&1wC84DUddVJBftj<4N2h$?kDK{VmWyFwZa!`!)$QrqORjuDH=47q1fX2rIVxlw?R zw36<0Tm_rN3h~8^Ha?rSd^^Ws!DQWM#9rtJ%VaSeEwskgtMK6sghLFbs8je-M{$qz z0iMsHMG0x4nv1wPuTNK@+%tr>%is@_apc|zLokR}u{wSpM~a!_4moIrgQ?0AI(#j9 z5#_n?Tbs@b$}jr^4th}Oe7gQ*Hw*fv2rq9x=Tx-J@w^B$|KH4ue891Wc46+=Xt( z)8CcK6ZeuKs|{cKM31u$kGzolskVyxWASl9Iq#6-Br1t?v*#RRQER2SP8;?Ij5w2z zubOL@x)x^jvzFUw3$dN-DEQ#gNe% z+C^btPDFkpMA~7&m4$a|4~yub(n+lTW+Nx<|hc3@x$Ow^V9s$AQrIo zac?gadU|_oCENAfVD;=HJuj@ za|ZAu6Vc(Hv?d;YUWGBFHWFGAz@+6~c|LFNw>`evFZnLRTI2)PoQF1X-E zdulO3ee=CgKfZDU9HBvlUL=Eg3=v#V-#%Of7EL5^VA~sp>^LBetZV-y^!{<~XHeZwn{crwhK<|F&DMo-rBK|0 zuDu~%tD?@NXKIOlCU$9>_nYr)bsN?HNvJvF6T;8DUGs0YD+C`l?vGj=Uu~XnI!6WO z2>TfkaN%Ho&+BwHO8xrR&0>LK0p{>!OZk){k9Nmet(wfIhy_pmIyuCL7A8s}Sj1w~ zgOn~-sbw<4&_>qD3Ju{#yam1ws6uGfb$@e5awt+|=W8n%IQE9|SAiW>rn&BV(rc!@ zEXG!ondfJesUB*DBZhf$b3F6WFMmeUosiqE3dxEGBQX`&kUU7J$rUkzKYNw5h7V(! zCia|koPG8zR*B3m5>geyJ~BeP=7ZpWG=i&jjh)#^xT#d71`DkgPPN#tw_vzuB3x}U z(%b*bf5I4!W~_k-IIDV!pL(Xbhl}cyum|zn4ie)kq%-TJ^N=en%XP%VF)cXFlZ-@~ zBl!x)vzu?k{?Z$C;|hz$m2T$>dP%ThnR4Lc2ZN$m5yWr6&-ZLFVO*&C zxCK5=XSIYV9)2`^Uk04CokYwsc(+4klLScqB!hXH1AA0jgrIvds*{B3UM`wDjWh0g zGNs4vx^x6g_%+dqqO?j8%|q_ZDb$de0wchn9;M;OIr%b)aDeK@=Wum8R5q<=2{sq?2m6+a{-&r$y+B7#J@>zWp{*@b% z#~RF@Vy2y&8V~h*y2V6$7MLtMEvJkkQZw4>>zbpqq0Ja9J8XYrx-{ONj@LutQ z*b9f180l3`Le9y(X^u%JI9x{ZC9*4b>d-AoHtz!N#zuEkR@g*c$K-E{oF8)2yuGNm zrYh8m`>(CPLou#Bmhoa8#GA}nPGWD}wd-$(BX+w=xiL=@qUWB`2~R*eAs*2!@WvDF z=Ca{0+Pb@mPO~(J;rB$QhP=piS7lXFxX;_LX7F9y+MI)+J2rxpv!($(6p^8SZo+&R zfdfj2VJb2C65%mH$A{{mSh|77Pna`aMr*Akl;#X>4#NtIG0QCKndmsi!E>t3+ zz2*<<9Gf;lUzqEkyF#r&;-#fG;~3)Js;cn1v)sud#{Qq}w5Nk<#z=Lu{n3ZND=|xj zTO3at_92Ftmd+!y$P9q^zPKe9k{OYLjw0&C?6-TT?pyRhlUg|?Ox*S{R2gm>ULtC6T_@X70mfk@;`#;}YsX?LM?{)4^p?30WWR87HtCtoN) zeVbzSDB~4KcEg`g^w9h68@Zk9%N9a5&PjaQ^6EwqyTciNb6NST&Lwl+He&dn*0%<6 zxZ}CU?b#kI=q(Ih0U`4rM5bM)N6g-LA-gG#6@?Xd_Zk(>073;F_d1}hLG{0B2_LMT zv7g?03H;fdKcfDS*=~0NKbAp=J9Zgx&c0OhW}e>ww{%k0)!Wz}cS*WUukYlXgnRCX z!nRwlS6+n2oZyx0Mc2l-?Wc48#(x}I(N5catyk*OpyHEdZPsjh;SfR z@>f8gwq*UlgQEDAu;z=+D;!g>Hex;uvu*QakXllo-jaw7H$IT=UJloZ7=DIR&<&bEb;JQyI_v$<#|yu zRpL<0jplWL_n}tCMoX*^n}8Q4##hbqk&Q37QnRrY>oaRseUl0u)M!)G(Zf&Cz@Av8 z)TC`NK2*)U3@7kg8 z^6=`w`GUt0Lx-E3?;4~5BF4u@GL^8}5l~h7eOmS7mihbKIQ#$~0Pei+X9%d=M2Hn1 zZi8ZK^B*~xelhk8IB~b`J3yZSzjP-}qW%uM%?Npy9U^Or_Q%zvoYl*(kpds{*S3dU zWCc}haAf;LZlef`d8E+j6?9b)3$D7#?)q9hF4rsu!-`fqj(#{Jqd=3pQ8iX>%NqV; z8CofL#Sc<0pXg}jB{XSVSR8?U7-5;1OiGDM8e75a@D;ke_L}2%y+N|Z z#<{b}c6{}Fozd^o)l+Z6LYywWQ~oiUIYQ*Ii5b&O1%7MrTB)gh00o^lT$OHciom-M z1C57inJBO#75?hTAn5*Ii`+<&09{M#+IKqxeDDq|?K9D_#CEVly@u_lQ`MK($TM&> zsG#2%D%bKt%{O2hj2_~;oL>TM#4$_wm}U2YH+e{;X=$iAqVY>w-m zsWNJOg2!ro2DJoM8F4S_JpyK0wKEqnq~}@}l;Rwk(b^NvNS18;Y(9lg2@#&?sg{c; z6Y_1ah*?$6Of?q{IwaIeeImoo6>wk)bod^lI7Gd&XkKsRm;F+uKW5Nv6KdRSC}6<9 z=;UD7J;uTK%Ttr$Lj#oPQX7+uy{YpuHKHY0(M$n36ViEs>MUq=ta6FVGk=M(HiId6 zK@XL#s#dCH85=~sZCST3*Eo440j?N&85-Wq ziPPAad~pAu%7;)~_rD;|!=}uB7Tl{S`yhO3*P5mEa8J_mrlvLR z?OKiYtyqZ>f@uu+;|wi1y4w6I;PSA}Cm*`eknW8Wu+sI}3_8l&Toc0u@@-1^bJTIE zI>2}@1&WIInK`Rn>=`5J-FMm$q0L?aEd{oZ+fE!BWPSw?K>|d)YdL_QAY*l*P~+02 zS?mtG>;9eH`Uz?C=561Au2hiz94W9SbQi971vz`@4koA*uoUUaJE)LHy}lyUUg@gS z>cg*8Q)FAr+t9^#aDWALJwhulpz{L?00!!h^z1{o?n7uQ&&X7&`t8v3gTfYS1-}pG z9HhASrUyBaEB|d_ZR{KnN>?vOm0G_;?>V*L$YfgfU6IlVHLY6tt%_5B$}F^J3*6>5 z)sO%(uc;TLf8y8v9A{$z`+ijc6rfnEEAM4fSTpAz_CXW@%(2TK4ml=t%d3N5W2;G@ z>n&%A>zgC64*GiN+nT3}83O2=v1<)HYV57Y{7%H$Bi_R1WE@VC5`7%`1<{j$^HS=V z@aFuycg!z5hBv>LUidB&&RdDHe^~r2<3M*6AitFp^xRK&k+y7htemoU&F_>K+SxdG zlWA`Yifp1#d5CDf!hFn%-T%QFKnk7R0{s@LX{1{n>PIkc$-f zK;_k>%@D|W^f_SkzFcdt{5`l1bByy(fUhz}(`S}0!S&ZgZ>T7XBH3TdRJ~9;O~zrp z-_4Qo#AOL%sMRXmwGy%;eaak8|U z&GF$8h0|gT<~HI=qehBSjcqec3ltbL6EY@Go~plKm-qO^=pVONJRwZqp%^Rn=CV1X zG?J5lFtzm#p<1OIK-F|hM1}nf_;@{~!6Kh_U^4SLKN82IbVYh_2QR=qC@%XY-8Hme zMkXfVSY~G4*yS#cXZ?BE>vsVxLpNk5JOB9S03Q0UR3aC9&OsL#xb51w*Z6;%oz~CX z%Bn2fSX%2X!fW0`&NQ^HdCO=(Erv(=(dS$2-@$GknYV7BE({1?yL{LRHus`ht_DB)@Hh|{42eq)<7YZ8by_d{Hs6Ei|7(q*i zH`UYC1sDP^jn$nmyN^?C!u= zBd5Q9XL$5Wy*W8JyIHU#R*-5%mm1`+FKYRxtZorJwHLY$UK7yqw{@zM^}&(mV)Wbh zbO~D5b6YSR1EQ{BQC;Q(r5)#6&-FYoRxmFq4l*s41wOKyCKWu&;Wg*3wVYlI4zp#p zzPV%J>8bt={Ty|s7>A_Z#YOV-vR6Z2@{gFdB2#xp1(@^=e4H2%r9=fPwdn3F+Fvh| zX<0Fo#Rl2ox%aDA?Vu|dd$zkKRmHPsJ1>vI6@*UG%$jStCxOx>3EBkFgNQ4ZnYeT< zF$0Z@_}n`}6#qexhRhLiB2hu@2RB8Z-_|wcI*hhbl`;Dv4JZ#RGq+EWra;kAZL)gn zUMOgzjS>R~1cjrHT6c4P#u3=QzN=NLyntCC!`Tg1wW)xNL|_x4XQvvC@iZRhPE6OmtB}9pcp!x(oKaI&z`dj5*k|IvmbeNC9&NPciPO?X&z1~@Z z_*>{B6VW58PtE(x=&fBe2aXZ`RW;6;n3RMJqXVW@9fN8@Zj6AAq;kn209usrB|1vZr4Zy^W+L?g*mOvqxx4) zb9{a*(hXI6q`t`$+y8QYzs#QXYE~w|0q()g`wO}q&)O*a4(&SL|qC-C7Gh51-s1jbr4bWf)A$6X8#fYOYRVz)cnzsm>8g;h|RrKU8AVYT=e#&+|=4})7N(xrz{L?%xwwtA{ zr}xDYgnAOx2hQzLW5RdyJ&+{iRO+4Eyc?9vwyLVG(1R_@HHQE+IC|O4*qXDc2kR8n z!tn+kJ#<+njx+ghY(;sEWfp{K*^Cq;mIoB15|5mR*et=%4l+nH{5g{|pJwT1i5O*4 zS_s^^CI6D_`C^`hsYoD#A#9TIS2Dm~G&kl)$amM&cA;@m*?`@2YX8j=+w;J}GcN_J zR?$xd(jN4@j_Y;@#dK8ROhnjjFEjEmy*I2r38{T3U_AOSMj9=uEGsM!rz$X8Rm1BM z4C0}?%j>UOgU3n)R$QgrK^s6O3mbX=srOq;WwM9;=ot(j^zq67z2r+`@}~P)*sQ{F zrw~hH*;M$Qg+juq(c`1i0hDDneVa^R<3o;rk%KCIC4S*==(9sPD)kDZ)g0`kk6!0S z5AW*JCsi^v?Wxj9QF37>0xq9L(TGWEB&|rCTo(zFx=vweW#p1e`-$fy_<6Y3*Z#LG zHamwPo^^_CK6nJ3Ooce8=WSI=W(?j-ji003K+1meTRe9jz9ir|q_2V;UQZmx^O@iE zcYV$Fa6BWw^~>%YxtNQ$KWS}uo?hi)E; zzhZcIAm`Qr8w(f5Zq;s8b@jYM;+^MEy(-GdZHVy(pA@Mi8q}xW2@34f!v4XwC_*+;rP`R~Q=gYUe; zBKoroObpz)y&5Gk;@)ly|2(p$46rIZls`*LO(EI&C!+LI&L6pNizc(iKBsG)ML0jJ zw_2Z`iTx8)_B7S_vTBS{Grxd|?_ujI$HKrXqO&^=EU|QE=2pLYUP5B#p`pJu0xu*B z&zq^5Xa-H`nRvyYGvG{_hN&@q5Hx0C#pHuqD%M{yq(T&r6Mm@0B}3G$W-!qq4=ZF*&gC|Fw4?g zkuvG>>gwsTT{W7TokW;r54OJBFeEImR_bCx4_!u>A?^9>=W96a7LDW_!`*HHK_(}A zj<+jYHI*FX(Sf^n*Yfsfot?@3H|Uzr?%^Eyz!n(XDCO?F=F4UFx8scd>E9pKfP>?m zZ1zTtndvlsCNDmD!2`=XVB{>9mSE-g){p%5pbs2*t@pY5-ivA&y0nbYgrZn_aZHM9 zb?SMNEy6#B*~sc~Xoeno!J-Z`+wTMs5y=G63h>DE{I@v z%|rBYzi7jIxsz}@5^)ct!(>wUwwMSsOd6~|@yn z0hWYxwx5c5V=2OjoKBe2|r0%$~QQHAqpT)u5?9mdr#*;Y3j&d~u0%}qF^cT{x)t5?j_^$AZ7G6GO6bA^4jatJq8mT-T?8yVKP znr@@W`NLcOF3~Jjk805jtPA^6D}-S({n=&PHOaxjK~}c!>%zp#Ce>)oUmpM?HmE(B zL<8Y$>iFE}Q_(|qr>^aimss#m;MK7L1s(L|`4>4vTw9NUX2Iy&Hvy4ux_BaLeK)x~ znl>H3F4o2VW?T&F$27&xJk#`A(Ca|fjs6+m4HXZu-{R*aQ)qXBl{tmz6Q&O^_8T9K!5W9HPJj}n>m_a8zounH)3`fo%8Sdc zUY8fmDE5wI+i8q2cdFuydmd6Aqu($=&py2;*+zE28pnJ6y31;%QXd%IgoLFF5}b^W zjwt=XEFf#>6e0AF9O?O$fplO_;gxDX6!$dAOED=NM7_p&Ijo1@m6DAi#8#V|to^UV6hKqVkX#j5;YqwkidFcu+2J5?bl0lraC?oON*^@dta>rZ_HjBHASzXW zk?!3>Yuoej%B!8CZ70U-o4ApLkM1O%snMolu+adAYM^=sU4EIXtyT`fxi23A7HZ8s zd}no>@;tY^+i~7ZX*;X)g_ht*GHxSg^FKHv{`UE_6mR}*YiBgj5c z_!+oO=WC?WUNbEf*f=*|4#ujuf7LfUKrEzDP|=i&vK^jVR>PfQe{= zZAg$pY7#7XDH&DKCO$)|otwHkXmd_3kzMFhEWlbBtMA3r3f&OG&hv>g3e^Ez^H{B3 zp*neKGovHm^1GzVj*-wkq%_m@xDn2}y4T&rLcl;|!v64TC1alvmit2vf=#@lQQ>GU z<&x~YRVlQP)GrWB>C;+%kvhk14ZWna)-nvOKVK+Oa|&KB{-@T`SdOfI>YwCz@P~`P zVtnsgdDwAogNTdJ{BsNPjnpgX_C`zvKD@aVU6Qh|kj)O&{&@{t{F#AAJZ3_E#^wsc z*9$pFeD2zhKh@L?jaYa&Y-&^q`25)wR&{&dGhpQrq%sCAjh)+sjbao88z0r5gWT-~ z5UCDf`Yc4`av3Xe3qo0iWJ(q$pQVbK&O;jnr$#R%|2rp`krgVn4do26KXJlABoyKm zvvjs?6e#4xm0cHAfW$<@%&xpFIA~L(sWL4EiP?m~z$8B*!Je)f->kmdIq+CC>U!>J ztw`$hmZt#d^N#*{;lLHff!fbQy#r+%LQ7g z{e-O@m)RS3>Evq|b@iu&EGVMlhYpPA3ya$K{}Ae@Bg!cA=AnN<)pl;IdG>)Tr;y3& zPu+|j%g^gVdv@;8dR%|rWtut*4wM-meep){@$sQ>BmA>?Zn%+U$*k@KJ zXYT8}w5IPG1t4{7D{6o6T;&t;5&Y2a4U;ilP|(|Ml6!WQwpSn8kDJc!X{%{FlRwa1 zO(d(}IlgPDm~`lz_An}&8+ollS(foO8$PK39o;B!@{Ksc^ZYv|?M7rNDAJ}VE^I8S zQKBw20OCi*TUny{h5OD%On;)bk+&e#gf%ozE7Fj4b~E_Wln9=l)$9knR;Wdn=^ZR< zy@kP2drNq@hex(3r~>e+=+N)VdOz1jINMLTJiYbY-!%E9@`|pBhfTFjyHo@J~iSXoSA1U zT>1^tT-e^^a$Jf#6p2Zi>T2xkz{YJRww$%!Z$i1=DDP5Bsg6)5d%L<3S}Le;btj*nL;C`XJO%4_`6Y?^TnrcF%j-&`rnMk^ z{4-k)Xmds9PpcFC>k{zZkJSd7W}IM8cs?%slGUMuZKhT=tf$;Du{TYx5Y4C~yV?-3 zqSxHo5)f9Ice2_ZDpcT5G|Knz(=mTcWZ8Yi`edqT{EjJ~q5{(K7R*ya9)ikoz5g=N zHt*P6lP=-?m#8N^SQ`BoaE&4&5579DPxhoKudwKgyudn?hu8cx`~$MKT|fJXvP=lH zS5(Obt?rQ8|L4Ymd;M+0Gy+(V&47K%NeQKqCStR{IeK{BzYI+sZ=A~o+XG>d8(PHn z8xsW05fgFJn+SqQEWv}(_~3y9=y2j|cz7!6zZ+hLS>D3eVI^D3vOY975ar|@SFO0c zD?&O^bsu*so$p$?en=50r)OUoD|6p1lD~3GrC=MAKLL=y!<5+7*sjx zPmmh}J^C+}As}l}r4xOBi1%eDDj1v5K6B6NR@wV+s(OWe;_|Uv%7DIM1TX`?)~cTF zi3o*%X9+7?Yp1$$Mf{zQ2k;L~%ibW0j^jw2^CJQ?xQyfhQ$_Us%|5%&?vA469ZS}7 zNdtI^af{J^i*5JSd0Erw4gQ8&CSSS+Y8qYR5tPYBZBp~8YvDOVy4S6-C?5Z77JU#Qyh|85lS0Q>L&jiy~Zk?Bg8b8 zNm6)3QAWuX{0ZgDNQGM)!a9(k3ZY1mS)at{d-aVG*KDG4tAx^ajRHWe=Y+yLGDoiXx-a98^}~H?A?_l-|<&%JU@{*4ZU51 z>V-W8M}8%7(tJr6_$V&^L*&npF%Mz%B^^+SmmZof_-T$QYyjZMhpF~xk)rulL-xqK z_VT0td@}mnKQd*?2W334vpe6ED;}}-`FHva8z2$Ay8sBJgnxcU>>y@_{=0v`$izt< z{r=DQuwmLJ^Hi7#5V2}FP|!nieEN+rtwwSCyAIO-IRiIQ2wddowE5#$2Mu=n_F$6! z*|Wr^{@R=NP~7H|IAuUz{S@eoz06SEW<%sXk4Ap9(YlOrsn7&;5CXL$P|~8O+^#+> zHC#|JNcpkT*a38p$Hm`yZ!p96qOi7h>9M_%WPV)1?rhWY4;%8re1h7}+$Z~p!#LoY z&_O690&HG5wsOt1t9lALELo2S2VhudHT;}fgG=peeH>!1CM*(+X6&cMdwm^4zOAWq zKOgdxH)-w8)tU`KWduFh4?q4fv~ap#Je14{cXasvq17r<_4MZ1>z?^su})-RMy@mZ02BdkAaJs$E#Pr#p85y?Kpq<}Ed zltU=VxXc~k8$j(k&~6B2k5G=D)3pL4PJgChWM@+qA5T~b*N;$nP;4D7KXHyM*_SPe zi0RFIvz$=MJuow_&&z8k4Kg-n{SKF(p`0q0>r`fh;ZAaz=P_qcn>j#3_zX{D7i#y& zBT^W&rChq(+7G<6FgaM?dZEBJ)&7=~J*v>SljxUeY9rY`&a_buKe;-GP>|a8&Y3R0 zw8MG4EI}>42S4W~A6bjb5Wcdbwgn&aG8wYJnpfA=nrQL9tiLM3c>0Q)=|bUc68;mB zIAnTO#JSIdvH)}J*F>aByGka#czUPz@TR~C^4vCe4dz@C+pl=^P*~$fe)EV_Hb~PL z4({tNJgSQN%Gy@@dHD~9b?EsT@wYJfd4y*(sw<@G0|jJ3`FQ-_*zB`r$XuCMKux22 zU=dTUT$P}nC7vgCh&XJjZpZ4FxW)RD`cQ*=yZl-OUkO?kHwK&}zBctUCgHXP^dFOu z3rhOtr+FfaKk@~($OyNC6@MlPK<4>%we&+7GWGo0)!|bQ%O{G#_2;?1Wz+JQO(LpNDB*S& z%*tP26&T8MHSvgfIUG8pUrHNO3z#E2k1iOM!hWP32-<&2zRYzA_&1~xHbh)E`*ec4 zhh0Vt;7w8x{V=89Smqjqf_`{EHjPVu$Ouh@_#-i($%|(fw1tB0!;D*3c#Bjg9i6w& zC@8lHAr3niy-Y;fOMG*5SO!!q5~kx@LiEWO80VldA=kzX-Tb26Co--p1odNbZ>*90 zBM?u1t2GIs*JFW;+V=eIL1+1iNCW9%^je#lXZg%cfo*-k0pX(Dk7C<`XC5E>T09hS zFz=uiC+6^)K7vM-xJS%7zQtUEXNQEMew#AMEXn+kM2$g%phI2c~Hz@Z@ zvStA7Ux56c7XpjJl9I}qD_ZAWpR`-|WE2{_3UIsp;P-iHIWVQ!DkjOSOW)NrKsxXw zHLCht1!LShirY_gOY76yrG3S1BD)wY`5`C9YzlcizMk9Wt2S{spwP^3@}n~BF$H&G z1BSX;hHs|=#OQ)$JVqs_Qa6U(LxDIy5TRtCzbPDSQ8EzXKAF zaTd_GE4@DI_@#qcl7sN%sBwM`qqB-}FKj5`&4<2-pYz(|+>z|`-e@C44RO!7cLK0+ zJCSNi!N3@%2V9qfgx7=?oVe)Fu$!fd{5$d#l;eWksY$+hM9L*2e8ppbKrKzAwifuXF2Bs|u{i+&{1NAIQ2Q&``2g+aD)&IHcG4;IOa^fe?y735 z5~MRRy$Hj4ki$P;tAgD=q!r?e;CN@u`pL+P%uIgCclvzzV5sGVTf{r6YE zp=mXB6=SJ%MDT54K}YkWu>Lga_n}v>UC`sF1%{n{4UqXKR{Pp3{DJ4VTbx%Kj8w~5 z&4du-s}9~A*xim%f?Cg!A6~p8;1$9yjJ(XPR#IJ4%%@!=V}#yZ=7vI0`)X!R)IXpq z+5aHGUERS6gbkQsy1))pL_@k}b>V5|u?aLKfu0ZBze&}m1}6A=Z$DNv&%b-N-$R&I zZMTc`2!)$6qg}<_g@#Azy^w6k^^aZ#wS4pZ$`sN>rtwkG;kB0@_vID`I=gY!XNq$6 zE|qiWiZhk&awHh^YaGq^#OJPTzI?<~2mSk-t{v3bwA%d)-IaKLNSb-~tX|i=HK97LNFVcqp#54oyV;4}J(-$= z+z+3uk=i)yr`MFz`-2y4OgB5>(33^;UbhKJx`PJq&i`gd@1V#=Mh9E#oWWjFTzMYQqgKWg^yIS+ zC%<&~ME(-z{eSD8?GD$I0cp~6f>eQWEvTD)Lu36yZL&YP#N?Oq9vZ%V>21+3avLH7 zK~{Ll?4vWrW|Yr&5j>5UID?DlT_!EdPce`Ad%-Ej)Y1BRJ~?t4|3oH)lNUj@hVr@p0@wrVWBo|?70(_30{Fj}fX?doXh32t)eP55(3n3$IqDG8|4dHh-%-k!4 zr2X!h)IUDk-It#~X+OT`-wnMgiAWUY3E z9GYWhDVM1E#)3x;ubI-0g2=nmF;w#WchbF|C_AulxZ4)enYMWJW32=Vm z(XlGoYCaHswyDe?_P;Pc@ICU!D!HDq_X2>uC8R zyJ@-Wea}{_<>IkuYOfH6I?K@dOqz6i9(=z;&))R5MQiWK>a!Ou-GecVkRHb`#pjSw z+gg$0K-c8x0Y;bK#PA|Y%5q1Ocz;P7T2p(^xJmDiYXE^PZ_4-oa;6ZpQT2?Vd?L*H zYYiiIQ)z&sSM7=g91Y#R&K2ZJg1Ay{tKALR{pC?2^eVd0BUMex76LtO(Mh(zk^6S%YZ*Ye?*H7}ru*6ZzFr{3w=Mjv5X z*D3^kjy{X2hW0rst^nZ=eL@q`1z}>WD*?A3mxR4lvD$3=#@qczq~Ennd0Y&$huc-R z^IfTj*VaZS0LM;HTtcaQ zQ_iMy(c1Kri8Xf=MjT@&@G_V*=f?)h44RS%QvT$Zm8bg6CMu6(^ zMfaTxDV9uUUWEt_%{MI6$f5{9oQPh^u21f+xMs?{JHWgiFlw=hgP)kX=ZX8Qn>7Ws zC?hl#2c6}mRRRoVFV_lx`TS7)7r+{|e z^O^OWDVy)`Hk)(RmgA=7UUQniUC_x}%V^&(!PhfdP<#lmIMD}f&}fVT5w`7TY@I>@ z=>3^7NfS|J?&4NGJaQO9!a&;i;BSkVYwju{#h?s(v?pvNR8s9)@t9wQFMi3zHwr** zwGsFw;Psx;d9J{d@{b+i?kZ^k>x{i&hfp3oI#JOM_SV+@fy_SW8p>+4=Msy{9m|b;terN=SS)3YF-c?vnfWOX2I_v)Hpg-00j%u>tqdSW~ za?6MoS&|#Lw4}^F>jP2g{U2NIY`z3d!I1q>14`?n=4-v0`Wu*ZV^&ewj7KPa_G!U$ z*`b#`l)D6Hg4WNZMUFJlp=oh$?-3z0#n$%i(0(44bJM8E@D}s$#$hMs#=LT9x0@QIP$gaw-f_u3 z95%Ed;5AOS?+r7Q+$VE+>SugzgW>QQy{`LjLX?$4!>Q>TChV+2x=c@N@!jm0t z$y0eLX5sd4D8VoQ(0yJ6)x>mze(`C;SMc%rzNDgB@Phz+)XB z`a}rHP(apuONblDfL(PTOvsR5LZKjrW3+jpZ|7x*VaR|k0CMj{2qN|M%>;YaH(=_# zXSvgbCt2vOM{Q~eIlL*{K#MK6GuY1r$a!b>`YK7kOsYLnt>Fpm7n+v<@5hpS>fMLA zuRAD0khwDsi>l8cFFkwI8(o?%@=Ap;`)`5aFlJ=4YxHnJp!zciH;f739)^wsbHQrM zeBy`AOnugH%0La40bjxeFzm+D)O(K1Xg_yD>XTNv?cMd zO^CV&yoS)zVzV3I;gKmL;3|FQySLAHknEMqDA{1?#1ps$anw?|%R zXn6jzuJ8Q~b$s$+&Ib9c;Y7p?i3vw+wdiO*I-%PgKl|gna%JX{F_v#0mqPPe&bc0I zu!p2tQ~vo`yFqt)$2=Z;9XX4Spb$sm zu4~@VbEnXI)4@4Kj$l(RsB+)avctEHl(aO>EB;Y|?NmRbKbA(wUXwRC3j}#4|5J;w zXTkXut|Z5Ho%SzR>}lp(+3)~fR}*8R$oKEl)&KG_!_fzR{n$QnOK_(yI}kEQB6BN26HG*) zny{$3e_x>k)L+tVB^<#TB`fimajcI$gXuW5MB_E+3kr;xYi0Jl+Z;9d^@RX}+VZ7r z4r&z|sYZBv^#1wGfOiFW(|kw2<=b~>NhSW9DtAH2?V7LTU!A0!W9?0%3-%{Im)^xF zMm`gz+Q!t+!z4{1P7eJo8S15!r~X`kCF@VCohssXY%Ug=KmTEVJ;(3FCaO1!O?+oT z175}s*h+-b+Sep_3c{$mS^$iuLu8DTzicLvTSNTHYx{isM{_eg@B?}0?MdbMZTDHX zETL`imf`BFJ3njKJ{x>-HwQoKYTc(qEnTpCeP0-(eer{2oBIigxb4{nh0?>#aUVOR zHFy>6EuTpcfo^M4(?`nFI+1kUG_o{0qr*UCH+UY`*Jc-SRq>UP9>;70%1dIQfHK)c zvEnXYcAu(5;FS0Q^pIk>#B;45B&tvL1q1mzNp)%y$zb-;BUa+5Miw#drAja%c^R!* z(P{JNk#9wq``4mjfQ#BOr`0I~mS{tZodyUvKNS{5>1Tt-bxh1|xPO-%Uf`bK>z zk1NHIIPM@V9pV)!Xu|6I`7cJ z!T+c@ZmT30ojLp#hQ{u1Xhas9J(Bf5;EdP|IOV$%5Ev{{a7_BqDG*)y+jVp9oi zocJd$O4o5O1Pq!PrIK|5p||c7d4dbOx&IpCSh$O1l&YCjy^P*=m#U9J5%6jvCa2L} zM!)g2vx;2PilmO?zUs!mject9^0+T!Nn3`$HABwQ+&bV<#PLD(>yLYf1?Z#-{fEgh zL9>Jo@yUnj0|FSFX9x1~|Flrc*mU3iR^5Z=`Wd}nRX=8wI8d+}oLzk9bJ?`91&V`> z6HJs;T9l&Ll8&9~?y5Tu6q?T5QYfG`#6kC@Po$YW45kZcxHkyrkm;l&&=#jA3X>|= z;H;P6m6zQK0W+VAoV9K5@No$huO^0P%p#l@i1w>1t4CkH#&&aY?sC1*>5W@wq(%roGO}UGW03ca zt|(I%WWkYsq(8z9k#Yce9GFy`CwpjKP~JI#Ka%ZvYZ;y__=f`ZU&iaY$>Gpi^l%6B zMXJF2i!Y+w&ORG`@sope++Rm@(ttarxDlknH5~ypZ`*TNIOwCx0trUyZ-5mjLvx?6 zzu}lJ|GTb_0>5Om_?Zj2ynHgcr};1{`@Kn+69?vVqpvEMi0Ey-+xaj*U8H!y_AI)zDO5)URHPhK1FMJ}Y$}f_?wj{^0rmssHl-0gN z5m~t8B<|L207g**xD;J^wB(12n__LxfS+>r_!g;f(xvzRrA&s2ufopRGSk*tVn8b# z4fDm%OK#`aou3T{#HcQN%&_&a-Apc;O%t>=o|$+nAC$H|7B^89@%0Z=Yc%q36}=Uz!z zQfSnVnJ~pwa$3$a-vq3<-CV!%svF0~SqjfbI*qlzS~H7)Iw8_S`_{<5wg?m>zAqmS zdjFg@txeYYdK@PX!4Dn*0IHGuZ6m1Z90Zzw!Q{(P+C3uez|tjoNMk*4P|6UoDVx5z%+ENvJLPMMQM+!rvI`tD;cJ1-k#3YOA8BRqiCk2S~R z|43*`0v^RCc3W2L!$ZqU9{ldC#9d}zN;&g3j$*?5#ZDxJ?iJgg9z!Vp^&6~_eGb*6 z`k7wQOUeDnlo>H?M3pi9x87Uel_;1*FQZ*Yl@Xeb8 zo9B=E)DoxDwbvIA$6f@dgL;Z*KOok=(t2wBKzlE&snkmy3)4vYRZo3SK~W?Dqra;UMFDXafQ9 zV~Bz926VSgJRe{BTk1!b^uQ}MYQ*;MH#S0pxRlC(EJL%vsZl0rsp%U=(naw$Ns0H(+L(x#Vc7cd ztLpcqPhg2i%6+Ggh;guaUi?!^w;hbpl|@>@sLN!@WnUnCM5Ievbgk=TVGDKf#f4W0R=php4&iY9Qb3X&^cE;*W4KX*6k&48Yh;FAK zhygaPdhJXavX-r%4VU(yxJnbmUeo`1uh%F=R|8FEQiiTgPnm)I+)P;Rj|jY9pB!`x zep_DW-rZ9Zjr5we?*X}q+oEdUL*)?3xcmr0Q=JUKG)Df`$$Me$8T=Y>1TwFy7Jk#r`_R*|EU2j+2=DN8g z=Zzn|+WVrJdc#k^B_m+b24{d?D5~gFhKOkzgl1I0P{;dc0Qyj!_sEnA%_Bi98%9Tm zzW;4HCf`?T4kGaL-w~Sn5>QvAdt#MIlQmRA`mS;|vn}NtZ`?!SXS)jC=9d$D6j)$l z3O;=*2IQq_uh05J&Im2Dxy*ELY%a3UQUB=GtP*?Z$Hwy@B~Gfau}CAwyL5Mp+xg?K z_>JGr&d;We0%<(eeXn+YW*iz*Qv?`rm0=Oku|Cl7hwr-kZGVG6UICD!0W7dWsgNp@ zOHq=RC8c6EnQ(#90Y%y3lFcAJzI3)It)$o5-NRWDfQs0E>u&~+GFa`2^O=~KR%{%^ z+-*JkIwK~9nwee_jwc@I3V8iOSp{=Wxea6Lao^V@fO2d-Xm|WHFR%8@+Xr_;sKPp>T>P5CGnC%jb-YIB-0SqUW(&WP)3NAh*bg@H-W>}Ep0 zfRoFmPm2Qn^0r<1rM3m|fWB|EB2aV`UfE%vWffM?ZcgTxD1z+f^)-vYN2w_BL?A<& zzVvHqpR)K! zi`cF2%4?6fC*G!IM^T`h1-grGGcekhl8&@`G?(5nc#sOb+8O>ACy)t{kV8m#qFuA~ z2EabN`~rX0@6W_=eCM;9!}y`s@3Z&(wGULWLfTK;*k@+Fh*MtKq@|)nRj5(wvq6{8 zK4t4hF8^{>f)jmH*8AhFVFPodfC1+WtX7Lqm9#j?gN=}@n_Q@rV=wniSqU;JZFc_Μjcmc1hnemFy5nzrSKH?wWNBDBn;u5 zRY|sVwv=1bPR6>D0bR?UWsl(}6>>~29%o+>`n77i0^&w4su$Vi_yo+~mB)SH2L@qE zl`nDMZA2{*k$gU)^q?*!qJ}sIQ5>{3zA_(qDVr;K%+H%P4fMQ_q7S=ZfsXb85mT#y z-DvD3Gr!~E4stSp4(hiYE|1LkK$S5s0mQ^e0vUVNZC`a143ImJFPmc7bgr~4NzIhzzXF{cvi zHX+7NvOihz5*9osis<8wLNf#KSa?&`F=JsRAVQI7l;e80RTwe2XG7EhJl?>_Dp*By zvC<{^ql5Hrs!X~2^Xa5LGG&H0{Mu+9nZdxw$319jc;FKn|GeZrJ*q7=S;QJM<_gE9 zcb_qh=3gcFIhLuYY>kMqRys`IRDfR{}o0)YE-fTaD zW99}Zrva$mRPqKplOWeSu9YniOG8W|&A8Sgkw7siwN2wU$+Wg#JUqP$8|#(PxmR?;P`1C1-1GM%el^x_VmVfb1>8!PvT*n4*;OfL<;BAk8|pr3rWxU%X%LENu~)7o(v+#VPd~|PM%~ZD zc^92OO{RT3dMF~{YJE5PI4#oYyBsyNA6yMhIM~z72*zUggk1WB!vBD&75QjD8|5Id z|J~6vVLZiyPQG$IRRx*7Z@?Dq`ixuI2C{!mYuDra7UtCaqUPfp zb@r|n-tMQ;^b}Y@;;|uNG(zsbt=>W5!I8zPpnVq}ug8<@EKfTejy2GK9jNqW*5Ri} zG1#g5E4tT#WmVIF|EPk{^S~n8vrQ%6NIY4}AK|Pe+pj_zogjNW-*qFx%_McR>;yR! zfc&jYjA?Me2!629`kqNJ!%jf{xwKLIe$Q!w}<5+iJQ|#UAQ~-w_<+xSc$F4LZB{}2d1p#RUkw&_EKe|I2iAgIVT@s_aTMOg zil^KUxSMtTw^$mD5{gNJ)lT0l#6NN;JY$50N{~4-{)dsL&oK-KNlsB+`O7{N@%QC8 zhn;rU{8hd+hrxtT8rHil6MwGWAv`Del$*Pcj*i12eH~)J&qT;Trhe%m%aC>jWXoXr zLiKTnL13V!^JmeOdvwZ7+)y+D)Z(`}8TRtnp=q$^j=b{_x5Q}psT7?kt}%P9Z;8)+ z$9%S0taTH0&rX~F-KtXx!*Wo%$1i=Q3wz#Sqqpe1^%!fMxU0wmnDf<~|FY@~t>=+& zLEp1z=g_iRqBb<;8w_qQ(Txf6Y@PTP=(6I88{}@7!$hP!pPJOVFcH)xtuiQ}WbAD9;Ag2nb9vJ+fv}F((#m~;=bTSpnMX0h@8dF)?o6iJ zyHxp%Q7#c`ECykrvdfLz80$@JT89*XgMq(mKMA`nhDqeuI{ zhJtiVN3U9k!HsSSsz0Jn%*se$`^yY|i#Ar*@ASi!Pe&j2lN!d- z8Q~ViGoza$_77YODVMLqzHjJ#E?f`L&%{Kq=~~J5P{L0k>5xjG(J}$iA`p$O51$`o_`h328!Cgtur!He-!yxvIE`fww|U5KbmS2i2``Og+e={J=CnaJek??G>$ zJfy|F*v05%2-fswBW2gUSQnAi&BrPCB|dkj3$yKJn-e&soKF8P{+o8u_I&Y!;LUrr z)5%=QjY$?y!&~opN}92kKM>RiV=+QZ?E zs{EGv0H2Yr%&YYs-e$lZMXMHZ!SGG?Tf!5R-KVf8w1NxC5#$S4MGX60u3McGp!!%B z9e^6ing23K@A|1G59&Ow#S)n|z#qaJw48c(kd7zbAXc7Ii3Mv`-*=6QU$8c{`9`FD z9;Md!Kto9lshe3YDjk@#?P|fwe>xg2dri4J;VSsRJE`3nUH`xZ&9YTIT-~0TfpAoY z1191rPt~6B#KW9IU-SO{ax)#0JTqoNB>MRelf1kAQ8I4*>Pv(g0!89Rk<%>R8Rzl6*9P9CL9R9WS<^iU}Fu!zejO-m{h>?`5Q+N00AsD!JNnFOT& z{haHqyE)q}BF9yF^nq+MzM|5rgp}Zqq^s(>K9bgXIBTPUBw0j`-Cnx9zh|C6sV3fq z{J{Pxt4A#4U=ybo?dHECny>!u?Y;LVj@vg~hv!dqdeD9fTZ@p1+#>gsbD|SdAO6sr z1dEAXTbGW_fo77_p3C*&^8$t_Gr^D3t1mY}HKvlKL$4PSp)0CrUa%6aO2qshQsNHd zx4oB#zk_ol9*YDJa-pN9;o5;BoRFQe9^PHk48Y)Fgt_pbycyJ*a;I5>bV<>3h5;NO zOiuv;JLBK|DB^Y}BfWRG#w+4R_R)lgoZ`xm6>OJl@g^;kd03)GUrcjAKV zOzW{d6?W|}8KOa@?`xGG{)k#?az%Hi*5?cM!CL2Ig7AuM z7ZVny*R#Up&>`jTqMs5t0@+E($=JX0$}lSKG|yD?uTA@CDKzsX>%TB|r+R=>Pn9f?h*-+jHGsRM`%m`C zq-Af+7{>7OJ6gtthl4_Sqr9a}%ESFfyMb4*o0$qV10v{~he0UgV+^_hOI`mE>3vF< z0<2|X7oGcnp?m@8i3mQT9E*Xzmdb5C5Q1yoXkv^Bx?I06?dP#(SJ~F5Y9{?Pwv}U) zK5ne{9teEACa^C$r;&VW|mo0?|9PRtB_j-7pn?>Qp zZV#mfCc`36*QvId0_t|$qIOFq-WvC*@(o%HL+Z#~A zE###mOR0Swa4}V0erx}ggQ+Jf2+lw!x=-T>ED(ffv_yQc*j2eMMY$}VH=uZXaFTu; z;xr5+A(Ry3XW5oGIwt=C)$rfcM{lViK}F}1|Lzl<-#qE1n+pdahn8)Sw0>Ge<^9u2 zxF&~5{v(+@m89Ud~VfX|R9?%j9+?R_@;T&jIrw1O4^Z-2#=$x>cGAO2EL zN6H6g-GSD6be&B?Das*^atZUWa~(`|dF(F8qw{^O5Jv&2AMsbr!T0Bs_;orsGqsbr z+!MbX&I9MfLoie`zHdg@T%^jYb0Gi1u*Qf{`Kc0-avuuZFTUWDnT=R;7~ElE2K112 zP@s$NoVnqZ=lvK*3~r83Zzh^M_%a7R5 z`@MS#V>ztpskwx4LS)=j%$=K%O`idyTtH|H5ck+SxIV&_`009UVtt|%tw|xBlP3l> z7vLget0v5p34G08qmn~@oMDA@Jc=06afo<;&9Q381~n4cRRrk+ z=fm_ITD*h0_^d%^YyMw1T?q`I+;VaG4OdZSff8i!AEyk`(mwu*eH+2@@YE2zI{!g7 zn6{J4*n=P#vt6B6LX>Jje`O=3x|(WDn6(#+5Q<3|D*8 z*$_+%#}lT&dJ$=*B$;`A(*laZ-$WA6UsZIxKlxAkGqhEVsrL1VVAy&|v&Vfo#ofNj zDcSiuenr!x!ZB=-&#iuzQhG86fKJj{y@7j4eD$&~m+$Y#Ymsm4o{A5|2!_fk!3n9N z@SqJcQI#XMlUYt*0*dqQr}oH@9zIkox3a#!Wn~Kd&tThBwWky72Tz(|bcYtJK0PeR zIl$$dO*q5tNKvhGjEOYV=FHI`;)z<%g-?f0IhD(Nee3&5hXtL#oS1b_R#kZ~BQ5bU zY-VO+`qQR~p`B>?uOsfY>G=>4vH5+m;&YtVlF={J5}w&Rx}^#B$FtSfEl;1KnBj5% zTOgu>CkH6EBM&RLG`$7a>|b?Tb55Jc`{eAtSC2~aC%5KaDd zqxwz2$){YOwFTD;mMpRHbb`X3MhGluGphx~=@~2Wk6Tu-SJW8K&gr9!T&V0JLXbf{ zLSb~K>e9X=e&?zA2hL%r;2`PVxwVXMmh0PMrUk%^UrWvU(j}fHM3=oLiQ1PgDAs}z zo`Ltz&x9z)-T1rxRAUE{+MPL{fnVr5YW0gJlm7N_hte{Lv&*gB@(=zSf@@yhfDiBY zY$uZ_$d(|_dOwe_llfMLujyK=&~0y!I^()#i)7NhGtEOq(M{H@sqJEp%U$Pf%-Fws zedE%sd%dJc`>;|q>Z*rX1U_W%0XVyo+{S$;!cLz0q)_x7Ub!{pI#wrNqLNVGa@Dsb_>=8Qg|t*f`fP$MPl-I`9LglrfHE_#`cQA{b*2>6|uKY`E|A@_U5CV%CLU0qpY|gr?gf2G-ge3 zu|EMuzb^q#ar96;=w4C(2+KMh&)8e9cb$ETg_lJQhVSjd7~FU`7ev5e;DD9J%2vh!H2m{ZFv8F3 zy-5m&%i#1BPQd)!R1H;mseCs31*x3@*#3|8;k4mcjBNjTg)y)EJ&=`_(R$h^|7t2v z%B*&)W0`H9e}VPzoBd^XumqQGy-Dc0j6;c}6l)XMzv042hWuk(gpC?47f}VXwrRw~ z^OSZ7#FVObJ!>>_kP$4yyy9`U3dfZ<2v;df1dT_@- z%O=Z1K4QkN8O&lU`CSj~ei`P*TfW^^+s{)rUA#Aig4E}ONT$SH~ z3sV6#x)v2QtD7r9!aw?vB1Lq%nO-=A0wcvB##ax}Va`$px@WQ43vfQlLv%6wqFJL= z7k1C6@D6aX;P$R3_S!T)HiSa_pyri@^&k)vvcD&{2i^jumA?b(Sv z(}T*Jt45VL?=EM3tmnu?U0Tb$Z3OT^U4dQVYGss8gR!!7M2taGW#r^!#f z=Xb+D;dVs|26Y>ZJX742!eev`a3XL_?77KkzvccsAIBuLmomW?!CL@}2hy|dp>W8o zzUBVYcG88}Z^aSrK7K&hd)~j1??#D_@H#V`0{HG<5CS6`ln>fgB3;~iCdg$B?01^{ z@7hR*B4K|Cn*}I7R#h6&WN#n3v}$>ADg)~Y2^@u7eH~vm2cRc**T+mZ=}QQTRfK@u z?bMj8y$Cf;6cm5#aAK3U63|3w*MAz!m^wS)Y4YnQua9#3z2kZ9_9upQ_08A-@&4FW zzq->VXbQRk)AD7@DsYFtW3+XRXm$=4!i1%*q9fa%1cWZ@=z@MBSnm7bNhdz@cK4P4 zVbSjd^V(lRldHQR>gl^a@DPxxmr8E13)^aI)7ZzInx6Z#B}ghj*ttOy@U+MJW)`^Y z^_efWk2Ks$Lc-O%9c@b5n!t)*IN4Lz@At25`6);ki~EY3^5veK{r?lawT=$LH6t=z zuVMI_*#7ga8m7R8l`N{TG@^9yTFZ$(GvH7(PSbWu{3VkqoeUMZx%vz^JOPFeiiz~F z^BMtX03~!~|EqC{C4J`%S#&kzq#6@Kk^3c}V@7I=9nB-qqjm{^Lnzb?Wim@W31QH? zx-MbF85%sMf`-D~*e;%{USJzJiht*6Oy!i0*>2Dg_eGDnZ14`O^i5B0w~pIw&SjP5 zx#ZF8Juk$~yz2v7i&JOdrD}@r)JF}K*~^@(b@`0fL-u~}3A9r(=NfFNFRLQ->2rqF zVEp!uJwEzdc)j6C;r%;L3oZccLZ`3Yx_edmFW1Z{RJx{jB$ZyK|7OCXM)-0sr!r}2r zLGhlr60&apAK=6kZZ(wnP3R=7mc**Rq<-2(u=~!6v`75NPlZt3^&ugJepU*(KKk`8 z(%;r67$)wG$?#g(4zfwltybUa#JNQ?2)(rvw3w8B{NBy_TyD};qtEy|Y+;5zCO>^} zNL`RD){cFM)8zcPF=XtGj6?ZAf+Q#`X{_++9Z08itBFdR6@Xrhzy}69)Xwon0QZ3D z`>>_|bm@W)ESonm-#Z28XCBV#Y0u@HP<;=wOGCO4FUs^J(1K*vJSR`(GeW8fw62Xp z!CJ7JGtQ^?8{4qk!oMAW@*I;}7TDX5sz~oPX*5v(O#QSOSEYBS8MGXd zO@BvRMI7qx+qG^R2eftk-0i>neJMJRVUTjI?WRI|9Gl8OgB!^6V&ZyY8P?SAIxs6jn`i+4>H4{3`A+XbypHuwoay>uA%yJe8?k%)eOLI3&7HI(YWB+1Sp3$u;vt zWxMQ`GPz8SE7AneNrnv^u+?PH%jys@68+uwRgbcjne)n;bgZ$cC-C`i0q8=`4Var# zc|b7ADX!*5=$BKF)%P{BaKR{6C}v^j)xM1b_njs82r(}Cqs&?YS^ zaSYk~_S=KrMVB@Ti)oj;OIIgU_2Z=|F=GA)3?JsTTQdxB0tA7YCq+`sx59HMwquV< zqzk$+ULs5;$}#Uj`RG%snZZ6NhKUX8S$|?SO3bwdoN3<=deN7du(4H)-WuwccYS(e zGqX!31AiyrzaN){pEjaDnLPb=sz+SYZb-$NF`bPHE+LzMISy5ef;dl!R$bM(LG;Jmk>Y8If}A{hEU2*088 zzivNh@?(xY!le_fKroomXh~ft$ge&!FjS+V5|(`( zq*_wI!@?s=?=#C1Ad5q9BkjhuRFR>0{FGRVk3gg>W=M2N;rU5Z(REd&(?t_*f5qU=vJmEtO)6Pv`m#e6>@K_0zK6UO%k&;L~$J z8B47>*8HwAqwhLn^|!rS2b2lax1qn8pW&lbu+7~YZR&pb{`}Th%L=KEsOK*c(7i`& z$A7cpR`Mq?t%{P^3GzBrB1HD6i%zO3#&-Ccev! zConQ|`{n$eQMs#4wKbk}-MH1>1FM~QN9+8?il~FXPiv`A9;FkJVJ03>OC{lyy<-*c z%#-pjxpAO1DK2EIFD4dmQ@~m2m9)ElS&Am1Sc(~V)pav6 zVNpA=dxf*-1%f;ayhqD$4Li(FvF%F^fVsj=>i?EN61wdLS*h9n5aEowfgNsy6`GYL zyq(VKm5MhQycaB4Df2Q*L$LuJXTq)EIpw2Eh@iDn80rt{!Mo!1oW`StiTQ}c8a|yv4tj2)1J#!q4A*(ORUHnc-@6sp&c6$My zOTsPFE}w@ir$FX#kg&h;L2iy(fX}ol+u{75l0=bg|35Fk{96-59A&jQ&{dCd#37Sz zxUaSO2cl%sMvb*bCeM&y4#21h$mPn>pzb!zmfxIhy6G=$*W!!{K*Wtkn5HNd#W!SOcfChnXw{-$TfNe4U~|i6u^5VwMvTuh1zGsdTfX-hZSU&@ zkp0I3DE1wX`H_J4hUzvkEk>`6-qGbq4WF0RtHzI`c-~H2eqU2Da&C3JcQ1&;_Vm22 zHdB=8$^TtBW3>4bFIJG|M%D(GT)4DS+&vDi8Ho0ykMxYEE~%d(wX!qVe&iq0rZ5Ba zr9&x2pDcuLP{{hPR#^68Nc_T%Q@?)7V0~ev)j55@I}f5|TeT(bXX)L{ng9|dKcfw> zo9HZaMif`JJuUMpxp(+SYk8YU!e)=R=xj-B-`@EJo=h`u+4`~#u!W8G{MUviH+$tk z&Ys=bKQ`niiV0X3l}wzS?I(%^HrWD8MLmuNY!@Imjkpn}`r{A_)o({o0nYByAXe}V zQ`({J9OySzgdAF5uDz=Ju!-4j(3{ZM9TusQ=>K%yJbe9~$PkZyU=n>`{E>U}B3up8 z67UWca>mCw!r*3KWH=$H9nsW&gQUC%1fiFV}bEGow^x>TeB` zA&$g(Xi-9Jjw-;46gh%++i1hIp)qkFlD``;#^Tp7B04V?0_wVdiv zMn8?)qS8Te9N!{qg1)oPO_ER58Dn>S*KZW=yWjdmAfEa+=#!n27KV!2hOFKX4Ls#n zUiwu0Rtl|q-yBI>2@Ua$&Cdy+HLbWba)wlv1KdLG zqo&S~x!bG(xC6jMR5gI2G|)i<@(X_c=YnYU{UxHx(m>beJI(Xw<(O&0vc6%q1!+@nOu5b+)h*hw45 z&qAE54VeY{-fu>o6v)lMx`#$t&|D|*TiO*QShex~Kt=_TnTWu7j)#pGc9mB#cSbYXf zKdI2~cRu4dxJqZ)br`>4c#VbuA3r8jVvxl2Xsh?SIHLtt^&E{RlxZ{4V7G=7 z($A9P@L1Ul%cKK9M@VhUBkQ%KFOSQm5J--Bo@-AT-7$a4o;6Uth+^(>(64_k_#PaEd>qQC`o^7xs^Ln{=3jE~CS25_fE;9EbaA=_ zZ6VBwZUCU+ky-4yxtIL-kgpD=It75me-~LKx@t?T* zQem}3)e?Gs(|=GxDMKyI{zkV$j5O_iq}3P9SU06|>>b?9j)4z9KpfU}QrgEc73Ut~ zd!3{t4RH1Wg9qpiC;t1?iEA22A(hO6M|t7@AQO5Of4~QMh`iWrYxMq*|Bi8a5W(^hrI!{JgTkS0;It z=RY>FUBm+s*-?$Ww=;z^I#6Fb-Xu*Nqi;UUKB`30A(u)I4L%uNYc#g6vv3*R_0wVA zLpDDO(Mo?QCx8~tavTqDOAS9ce|eq_fPGe*EW3g+7L}s!v>0dS{zjmaY?U@leV=sO zd-}usQEWH%3N6o&bEqUensI~!z$>Hw2=Ez2c1!?kcRjKK^_n@q|Es7s6DF%!i>|J2 zTPU_W?Wi0FIia-ELQS15QNkYwUmu$#(ffI}8gZ`eje%cSWJ!pfOst3=%QUd`oGR#l zv5XJW%!WunwzE?#lKWCIN>7@eCxRMFy{;&*Sag!VeT7}g9wrtDJ~X^e%^9)oAmp5u zkzGwo?JrDM+I^pwJ*EAQ8i(K$c0(cG(yXSI5_WXfbvW92NekP zNy#w*n0rXN15v4i&Q|K?UxATn4B=Y25(~;0Y!Ly>dokY~rqtL@zS6_1Qn>mK8^T81Kj0WRG=D4(}a+2fp&#Our z%-BVMn!&)Pi62T9esm8*AhfUQ_M9O0-`xzeWt`ib!Vb=E59g&#f1l?c=~p{>uex7x z|I|$ow>SL^@I)%fy_9Fe?IQC5d^Ox#hg88N4~xl!Ys?%f>Yk~z@WgeMTut1M@1*snkG{H@SSTfyMIhL6v~_N*S|qRi zm}->Mg45xy%%d`#*2QX=GIeE3@0T~Bu81<%u_n=|hlvS-EDLy;>$tdm9AoPs@%;kn z6(AxA$O{#k)%vl@UPefK#(n_F8wAKoEs;Qd$#*AqlrPGLzrkEB#o(sv_ziXw0zXqd zQwe?A-g6Hdic26988tuCGsa()s9|i^{?EK^5?IOE0P{X=3gEgsRc|+zxi2%>@J_RR zl8H(_{jnh_MOq_)tbB?}Wr^!;T>cR?-efKZ$OH}#?ewusNH}ke{s$ctojO~xC8N;1&)HjZK(hJ3xYmAMchkj z)A5VBsB=_D9BkGtk|o*B_P$_KE%zgRlX;uUDT&f*U@3$A#hF|;C`L0!aM1L|R2y&bOMU zpu1lQl8&|KpIyY~-6c)Zd-kZwn^73UM20Y;?ngv)_~AV*+d%u03OrGjz#evLKRc!p@XC{OibD44m)*~-ucYf z;gVnUgko<_$`ccZJtGn-swhV-O8`cZsD+ef=LbD(7L!+xszi6b{vUeNGaluv0O8MsS`2 z%tP?s)1E@+-#-|!Ecwn4JOaWoKpN{I#YcxVA({(czd!5fh6=fKnGE0b_#H`+&bA0= z_3eqZZ?Fsd83BikB+|gj;G!GA@xMopPuD0+67pO3N*{zgDsFroxkz_j9;#x*`JU1D zG{YE2?^d42gI=ASQ(fcR{sbBSD9gUVxGABq@vV+wNkflRCqC%1`kJjyz`g1J?nO;N z%&OcxWWeX<)f`IHJKsvx3vqGrCJmqV%jvsUkjkqh!yHHfJWLf$nWm?3IBsc?8Cl{) zX30=hVK0}zkz?tt1Dn!-c>R3d8Jfxd;FcgF!YXR+BGXYD-iiArY?rNjo<8t_SEH``nnn`*%dj72jPSM?OZ^;3lSKCb` zlw<8@&Q(VhDR;yz^D0k2;xar;#E^UlzJjnEx_rLBY7F56w>6Vl2-=Um{K62qxa6er zdH&=pHIPG#ik*jqqU)i=tQ~#kz*|78+y2`x(@c4b11f-KZP`BhdrX2eAY$*GfpR3H$%Qru{rXLgA7LUg=pq8>GYT4Pz!ZYQ zp%B@bG;?#@xo5g*Bpemnl@4ZJ@(iq7fAlN9&~Y;G^+J(-eJc2N6!`g7pr5dN z-sA1g3I71EZ*$|b8-*>f*?%HUiYYa?PD|Id(@fF#Y1~BA0GS4P^R>KlfN#NT>)orm zQ~E5Q50LZ?Xy81A)`<7_{q*Q-Ufsucj=*B5RTYF9@!znrl591k}#%bQJL6KQ|mC#5ms?MYF4eKH zC5YhBx^&c`MDTJBT-r}1qbBZ-$dwXkINC**Y+`e)HZ&JiQLOO{3?ggc0kk=c3u=h< z5=WzqUv>JIhla#EhRML+IW1DeY4fU?0|Gt7d(>>`Mi`mz$WZoVaD53NbTR*>Q~_ zve@K&p|K|9*jRe{3PtS?MQ`v~-+girQTjwf22=^DV-@3XCgcRIu@e1HrwW(5M?_31(?C6x?;pYe(dg0=&Nbt4>Y2813 zGV#~8QS#;nIdKBHs{&?CtD69U(3ebsk*jyfBlmr?tby0V(h^NC4Z{&^lK;8;`hjIs zzHYv(2BOox+m8yA4`aK?N_lBQ73Ng-(%*8cDZxJ9K!-pLXoyQ4iQ4k_g{&r!$s?10 zFjWV$)fM|+ZC}C4dKgS0c@Cza^HMpn@)(|_|Jau%Pv7>iwG3$$Lb7UvW3er>oZW8+ zQ}M&l{OXEG5@PKdqWEO$x%u41S~K$HVZg^l z+dSk~?9@b$k+tn74=)QgUPJOBwB9c!*EelIgS^o_*Z)J@6M-&WSHuNOb=d2j%n-rN zb0Ni(sulMWb->al><#>M8G7>GIyqpRE4?N4`WH=}*##;c9XF5G&TDyPb|&*X?5{|K zDx)Fb^9*Q7=o}kfXx4lK)XqfVb?_Gd+yOKPn{TG7>vHU3J{o00mcYvCmE5cjnb!5O z-iu9+qCOmn@b{6c;_0v|cK(vt1BULMi8im}J`WD0pyB9Nb;s=^=o}aooM2}l`-}Of z=i$mxA5fd(Ek(uD?*E~btI3^1Rd*v_13?%4OE=wC7kX4Hi}pa*SdOSR*&D{6lQ||D zf`(9X6@!doHI(=TNAs6$24+2q<5FaSCuv%l&7*ZH#|J<*X!Cg&=OS1MGy~B+6?imS zHt0ne83J+yMvq8npZ>_jiPk!UBx~HL!t++&nXJE%Z>`K>L^V$`gBrOw_JQi*F<#id8WLH0C$I zE;2BuEJEj=gT>GB=laxHjn)6<#GrV)?2!z#o_w-?hntG)@+;(_-ECW26#upe#)YYi zP;RXuhaF~_?lz))gMxIL3IP;Ag2leR<0&7hIG;f+ec=20 z8QFFd6UZ8gOJ#>g#-WE$z;zqX6ISQwaJ1rh2BmIsdJxz!G2#!D=Y&&%)AW9sutv%$ z4i?ogv(D(LBa9TFPdh_^;c1A+)EkP)=dUDYhhty(a8J4!si;Pg;$>deLzPDoWo9|usEhC3WKzecS)Mw{f0*Bn3%fa-}Tef`n6PKmNPlfBH zEt(&t!}yjNE<&PCs8l#jfr?@XDG4Jf-$?BxTxH5Aet#ZOlE>(n+`_1;gJVL1CmZN~ ze)!ww!pbM0eVfilV>-fW3S@Qfde_bdp)^k)Ul^RjMCSIuvTA&*OK<-uWN>EQc%W07 zXSSPIZ<1K7A4j~i)XMZ8!|;dvoiAXn-EZ!wj@~YR6aQ#!%p1}E9d49!?Gtg^qILsI zy*!(5s+2@T_pY|OZENIJe0gV^*a>*aLM1BFZfjBz_Sb?BIFD(E`ylU`Gi`T`%zH+% z?u);S6<_Ew0@_H1W;^*`Nys(`Sg6i2#(-C$RZ{|a)9@VS`pQR}_L zpsEfdi#2U_yov3)p4S~cZZH{&wbgt)m2hkh$Clr%oUMr4^{z^+mG82 zNEuZvNUJB}&7x`VEe#tXBI-IE$2YdMkMPMdBaS4iW0YyFE2ADELsI!02twP$k5iqJ zmz-u`^y_2&fye8_L^nEpsHFYp{_7CY$(D}IQpbnuU-5o2!{P+(YIb2-&gWnQc#P0+ zG6;udf194-AO7J2<0?mk&ZU=#Ea?!e`B1Nsu5>p?&C>76yb3}erO4iM@sn(gvB#>@;Ge?|-8_ZQM*edBT)z?d)SHDD|8A^W|$!h^G$9eI8gr6{0*WC6eBG^vS z@w1}~I4>tk!90M9ki;MhviJt-o2vgLCJk#rH_iU}GVH6MeVg|mXg-|YHC2D+hvter zvCK{+F~hskYef&X3vC&ed8tRicSgUFEWC9{JZPp)X^_E<2|jMKLWXty9*;wc)MEn zsQl%!2?@ulz)hCwbb06z6PFFlR&#^H=7>N{ehtAB5ya(x{-pAC5YtWz4@OkTUq)o^ z7(@xOzXWClG*C_!A8}GfmY^zBDARXrrw~J+9^S(C7i8EVT?;f&*&lkGNBui#7{|S0 z$1RZ6F?Fg8t$t;vO$xloufbLK;}XU&$b9weNxl$epd!_&9-*D|$pP`kl57DJ=u`Lc z<3+q`@_P%pz>CAA)~Q5}_I$YWV?a;AS1QD{!+9$}%KD8(kQ0rizrcjpDm4CA6M2o&^_;D&SB{ z1=1(j-f_{DZP!J&EmN4OF;dGV@;{mJL|t%l7uH zELtsU{rhH@zX?J#dL_oQ_Lg6m&iUm92wRyYMFxi=3uB%R|JmQu_?guZ9J+h~Uh~jL zCMsomn#Ei)?=viK4h$(UihyEE?n^7ZbX$9-a+AjJfOR)Jl$7rD;V2Wgk3jC12rq5=%|m1Z~LT zv`!T%amX}-E^3Tr)UMX+ZR=>oKIXrsj|d7wvY2dajwchnQs?_Bs@(21Xvp|LSr6Ff zH8ItT6-+m(S`U^fNLI8~teQDMGBO|r2Sxc8li?~~QdqAKRgb`*3!^NDxEPsB5ja9z z8NZ&qMKbj+_z$%4n717mOZN|9=Dy|O8K^V9T8*GIH!i(Suw#GaMJY)z#{zI5(NUR9 zn{{CL06_B?4IXFU2Zlkqy=&&}xqFd>^i|tym!~i+vZ~(5=e_GXqD}{Vivqty#*Y_J zmrD$5Nd39nDCLD*?=PZrFqy!|)|fQbUuf=W8^85HYo4_Kf>yskC&+6GWl)aVz0fu- z;{+P#>z9B}Ge3rj;5RU=J!>yMFhIVzUiY!r&z^kwIwXr&{~?uT>#e%UwQU2|n*1*M zXe)9TF$Frwy%AiGaCxs%K77Yp4T%)qh?e)dK`y>!q>y?))*Wp0Jv#!*az1rSJ7Y|^ zNVxVyjewm2-?9*gyF1c`A?rmd7NLFaG+VAtWNtS^3B4ECK`Ksf9sBTrELgKcZ)t$H*W_g+Jx3G$n{HM;?dnvrSEM&v{~?zzSwX5L?kr5%mj9 zj7Axf!BUI*Gd-onKtci1x0@FOzCu?qiCLfvx023oW!&4>@Lb83zxX>%#t&j1d+I26 zi$N>xdz22)y=TKgv9Nhlqc9lqme>z#@b<%-HI!3&>iknFl~NYr8_Kd}o9KSw^ILlu z#3u7iUoH`v15W>J|F2`M;sZZBVt(0W^Iy2)&*%TX^GWYapM}W&q~JF8iLI>M{{oh|Y z7Gdl?Zbq+?SFz|upbvIZpx!3&$gD(K!#ya!wS>)+?}KC`pvmWFU68*s?OXRbC9xLQ zdzhluaKa4HS;#Y~FVS=|t8yCG1~VYGj~0Hr2G2X1L(HBc4<|HnkSwVtGj_?eF`K+l z)JHQ7`hS#%F|w^9f3`lqTv#{Lc;Skp%}aQXBEMy{%y&|7XWcM{9JbDcHvj2p#7XU@ z6Zi)`d;%R(!(`$+TvpfaneIpU?2yb7$+GxZu_AEZg;fYrzdh;uw|5}~1BKSCR!mIMS5@PZQ_E}%ALazQg zJ zk|N|4UB6Pt79#(Ny)hwC@Gu?ulg$@RY?H9d*r3ll+}0}#b$1U#FL>X^GV4CywI$ii zCD2(!0Y7I^*?B`Y_kz||9me}EEXY#sDc_9_c3{~PsMaF%Ng^$ zdHgdHF4;=BB)<3yN~a8)j=>U%jIUyNaK&8HFNU zlxar3YD1`>HU3f$CZmvgR&ML2-bHTt;wSO5#LQo5S`t$U1K1mZaEstMxTgDo$JUMs zU3-r?brIhK|K7Nyfz{TwLhqCGINy(K#Wf=bBs6><^7k43gakHZm?jMo4qQxsW=tk^ ze;~eeiSduPw-06oSQBWiYdaJWBdja}X^>G!zV^fMJwI>S4OR4QxB<_l3r_|lAFP7f zJ%nE8Z2A_lj)|pYyah7*l?(8!%_@La7=V)_qCXI|F;sokd!f#`sJB{Zm$rs|o>V0MwB}q*0 zXWk zOCETEd6+(j;``l*f{7vitb;C1%7By$;v9SHE9^#KX&y#NYEj1I;6+7d6J>nsb1iKP z48U2{nmBFwhu^b9Vxz=nZlOY>b>@O!HpkF^fWhGvWj?;c!3DF$uf8H3wH`L*I-B%y zh<|DeUMLvqx*@)1C=&dL{J4WC zv0=^n%5cOG!N3OR4$OstUei`+;$V&Ja=0au_YA>o%q(==gXM%iyg~b2E+F6D*8NOkk?rtPTcY~5Lx*H{=OGG3^ zLSm$Jcj#zP1f)|&r%Ec_-7rRuZSVg7 z+r_%`rztkx=HZh(!K3Evc3!|m5VNo<b=wR#JGIzM$WZfdhl|?=o&iKJ&z! zh*qMH6%mR*5~{FZ*!2z}&j|%n6HY=e6EoP$L5aXl>Y%4mMgUxJ)QTB5RuQ#n?;9E)ArP-@Ie zfSjPs$CzsOCL?-TW$O{BuVvpOjT_j34Kc6TpnL@qe5XRscF6q(y1v2w zz8a1`k>+HIkos4E6U<}xQ0pvp|zw>mu4$?7v2;b&qy2%`=T*Ydp4g_ zQUmg;m2F<}8dI;hIwX~(kA$J-q7``w#_%q!`oi^bxd1s7v-^KA<<2$;AP~m~g@VLu zZ+1LBqRUClV>`W%N_s0{%e8gx^L3@b&<8*HfA}2}WujY~o6C?q#q^izNVOrSVu3BN z>me|8l}_>^z$vf!tL6jkA-HD1^{S5K!go<=ZPr*8E;tI&E-OY78V09dkyeeRsA6mN zp+fGZJ(hk+@H@Rl{d06WXW9B7#I#nlqXiQUoYhxEt*pj z5c$^MtG|Q%vZq0cp)Rn8HmZkyj=mpDP0yAg!6PXjPM2EV>*fkk-d}8PeM9E{aDslT zsWS<&nU*fC9N!>chsd<99n613NY+uoohW{#0g#t6vYj10q(XD(RpAEuXjY1F?E>MQ zS;?WFp$}oO;A+fZ;;*3n$(HuAxb;{ipasKvr{rOAl;1pg>Zh_t`1Nu91!jA0{4L)H zd4wNLR7ZMLel%Mxs2Med~9Y|Hc%B=(Mh*|+nV7bC}X5WyR zVgX2&aQXN^@@Zv!=7YdFj>GpDkIow#HpZ*oRg}A8?tg@6W7ol+Ha_c=XWo_sGHZq( zsE{ec2+b8tv=Hh$|LuDEOl`2G|EK1PhZgJAGC;lY!jUZ(L?zIM=XD;a?R~Y#(BfJCgkyTsOztPnaW6-PM zU>Q}Im8zlXk7YyIJZ;AJyM_pSc$IMayjeB=oL31F|DNZ`+J^xsOW5G%45-a>RUP$a z^p0u{+aTqyIUE22S+u`u0p@9aIGH?2%FkztDV@3-`8TGS?Fxz(GDPa3U)@5eO{)?| zi<@u_ui#>D*6<;&tB;x#?2wG-hT1$684B)KaHa%`gL7Jr6OZJtMEz9oNXy%$MsGSf z=So`&n)k}=9d`^I&FrR%3NR5;f{KDgLPy>yS6OP?<8Z}GbiQSj=ic|lvuH1t^SbuO z1T~`b9(Y|Bs6Q}Z_ocbJdOm%=bHMpR@alYu-1xP^m7br_u>n%P)oWn9^}|^dxneJQ z=kiL4o#Gse|AKkRHr^R85ZSQY9@NwYSk_%;2`K792N$h;B4t`VZCn2&`DwZR%I6Qa zs8r}-Mea%ht>9x!vhySil?&-U^Mq{s->!OrMi5|keBzw1sfrKP@?>{Ch*K}l7j4k` zQKZB}{^M~YI#tm1x3cB`)~O((A5pvzM6laGJt+c_#rR@?{DEjSUGaY1)j;UJoxd{9 zW8Y{PZ?`pJJ`A`xyMM6^Msk@Nt}3w6fMJr!>V>$j zfA9g4YHB494DB#eZ0sm$S5fU!hb;gqK0gt>{3y@;;_2ajF-^oA($d`GObNWdf29{H zNA}NKz0&6sfg~=;oEXc9jsJTbLftZt~jb3LZZp;tMp|Za}7SH9PlVJ*T&Cx#hJ%fPyubl5fw-M{-OF^dMSj} zXEYd`BVv1$o1{o9b-5nov7QCFg zKagadd`))fv(NKj|B=YZR=$AU9m61r_zFLTD}2zCY1l2J|OP;oV=Ig1TEx#Lp5F0#%xp+`cs>t znq@_*H;1n0L>D_LZ<(6yoAGU{q#zun`ZzdNEGDLG@-L7cV<%gqF*g%^8`+5iw(QI?H?0 zS&r(1pC4=X|24PU#Up#NBXWNW?nH<;8O&E3bKd^5^6D}K!u1>^#v31ZTOb>;b0oK! z;77H~mfxt68leAjtH=A_-e#8*qL7ol-{5*(YT!Dw1CDg7>YVKXeKYGI9sc%79u!su z|93gk1x7vYFA%Mce!j|l4=Z{5{zSZBh3(5a5EU$ayW4T534BOjcCWX$RWA@*OtN8n z2YYf+hUIyOxjAwrXopA$ga`h8J8xhA>5T1Hr}Vq2saorf@zbW;wF5tCGwd!DCcKOE zSh~GF{RVFEyQD%Goz2}lho()HR{LzG(#gR$qv-MwdMIqI&z_ay%(uWx)wz_e zarYSnIeY6d1_myNCILznf9*g_pv+L*82R~!)d=Fm;1*4EI$@GK{SNxemd1;cVh`6i zFAjR!hq0?0#rmw5F^_wn!$8$#7LZ6g]V?(uY}ckBGeLM6*z(s6#}`EF)`DrA8h>>|??ERMI_ zFg>%x9#O!ALy84cG2MoKDJ9}VjgKBa%K@n}O3|lBK|2rQTb9e!dO0Irw}J;ta)z@= zVOUy40?ys5O<)O=!tQMShkN|Sv91DG6e)(`%TrNyVe-S^zdMJJGz34f z=R$d2Uo22F`*@TVNWt(fJW!^0!YPs`2?D?de-58wA_#Yc@0YWCRFdKm?u+)4aoO3= zwB4}Y^4U*iX9|3NwD$`CqPEZz&~v=Z!-BiM zkFOFos>x+$+Log3h?dZzK+|u~acvvyR9vWXfJuztSA>Uw-1}S>`n2I|uVQ}?BZ?UZ zoQ&J_Q+(r>@?ANyOoHYJMdL&jLw)L;8uf3bg|bIk2;1WziFGx?dL_N79O=Umc~R6I z^msW!jg&13<;NQeQrN1Hff{4)YmuT;H9Aqf*qLHKdR}Bo6R%qz7yy6PxJPnAD<|_r z9++R)b@~zE{VQa&ogQ%}!0QOA`m3Kt=%Rohbuy_JI{Z0xEW2cB3P<37WdXGR`ck?_ zhRxd0=Do?Wa_Cz``@GcdCE^OcpT2V;%2d{qZ>@w9V%r5?VxEOb-M`<`%4};ScSl7c zBlt&rG_iU84|w#8s<*Mf;v$W6itIOx%TIWt1M{EE7>+D3@Tvr^>=+qv1U}>rm~}ed zQu5BD&Cr{2PFp1up*cQO{z}kMl~9y&#K;G#b7no0S>#i8hD%@=R!WlU9odSwI=t-X z(2X=}qI}>xC-!^r`#&99G1Q`cU-~{VMMJ-&KuyTX*VA`#*X&W-efTa@_^pbR&vt7H z-bJ;y+jn9^jo9xek^G zdSy(~Y*65s=+fQbxIKw%wdW&DkYbO}e~u{5WRliX!~v^uf_(LlksRe0yWJ?hm&wb-37I_dHp3vZ?V3n?;v{iWE6erLd3%NH36OILXsd; zI=6$-BB(8F;NyweJ+fLh3&AHVU$MiyFWr0Tdgnobe_0QH z#~$iuvRoY5Qo?w>g`4CC_2yH{aInoHm2;lS_8Ul2-Ie8tK05IhK!*LR*CV|(?0ie^ z_&9OI7-y8JUp}VGc_Fb*@P^-N=)UZv7jVpq>fqo#Ro_gTmFhLJ2}N~K8IMea(;#d* zi%M6=S3=)a{Ai>t1rRag6`ha@W}fy|zTtItkRLaMM+zQQsCj; z`PGy_HQ+Q4Pbjz;7aCcHA*Kj>A!`h$y z!Cy3%rWmuKQvXSQ5z(mKsWsz`CP{V+&`atGhEtEq&E)&z-zE)CxSG`tHia!sefG{E zf{j)DC~xL>pP%Qt{;7`*rE&768EIlmeH-Zo>iC$SB6u{}Ys~K~UThla=)IF;CY$@> zE0WT57iOqqrkl*8>enY>!I5aZsQ~Vf%~B4cvH%PfT}`JTH}GBEhZM>(3-EAPW8lhl zHzSVsLM;aCK=`+D-qCKxu5tkd3OjG3ld;TI*V}`6hgW z6*e1$svH*46+2+vr44j{Z79z^8uvuE3D{c8oky{Qam_Ku&I)FD07Mg;X@LLAXJQ|k z&mqZK{GCh;SZiBo1P_C{SEiWSi2!4D5oS^vrE)xnmS~zWVBuc5Vubyo^D4K7yUr(yEzncsfP8Ls#VIAY!jU zv!SfYloh3f(H?<~cX&6^jyT2$u|;^2@)SR1#3~n{m$)2NoY_~QY8tRc;*&jzV_Do_ z)k@2NfdWHvvfFT|*|F8uvszRi08-u*bk-#PllE2kyl2(*Y@jpm5_gq?z)TS&FKlu@ z+Mf0Z?99y*pX|qH`ao;4==31ux&iJ9dANV5??Hq{9F!SqX~IQM%YPQV!Ili2lHqe@ zER7|Sb@baKFCe5CWvkz8{FtXIfU%?YL}bF0F4SA?m)^~vNCF%~wp=ByUa`V7-3#_u zjG6FFsKZn1aayPZJH|FkcYKxN0k6i+<3S`rd1?@$V|oCsKYhdRv^`r%e8G@Z1sChH3uEKF+g| z9IbV#2;1~!rmm;$*io4>Lc&aQ?)Y|z!|h;yue>S$gAg~mi)9(2V{^z|_Hxn~+4~)0 z(A(ngRFtBVg*)Qb?*G33lbt{(R{9Qjv7V*KK(%grMR__Vf9USj*ggN{fq0aZ`t!S4 zzViLP$eAJP>Ra%sXuTVpN1g#)BtkrwnkL`0jsL5)5|4+XR+_)g&TZZuImDI_l%MNv z?dm>gFK!T)VB|V0B~)-z5PIwHptX4%>XV)#Mg&_N+QWossN{Ldb@}>QG=p#{-N=cn zSPEm2ePbX70*Tk6ZlBBi3+b!}DO;jvhgek+f3Yz|l5h8J9OEaf@zBAeGoMItmgeLFKOW49C}rhnF>nmo z9e|=0!#A{_m!X?PuHcxz<=*f8it_(rsR!2Yd?!%!uei}1nvTkpt@1lO*_$j&o*}^k zY#vdWdMY6Z6+u#ST?E3n1x!lcP8VUT{cslAY?is=`hqLRA>J9D@%C$KRy;~b}_`NAA3iww$ z7BX4F7n6ObVKTl^XR_Z3$Z{)hABk7&?!H*}RU$oN$eqH*~vtOQY!)p?V6Q=UpHUH<; z+P7v9IS}^a*ms=YlQ6ZA7#9~X)Ct%b^5fu2t95z6v{5OP$dHU=r`dR0hzf5DdEg^< zE~ENejlh;(9xK{U*4R1)oi33sI7pfEnh04pD^Vj=eeypBoz`z^XZ^t=&AU%+Wj?CA#l~o4|B6dwE!kNyYY*qK&y)VE z>luemIt?7O55xsQPwO=UvgjXySHIr(Ah46F0OPV~WR)d$bX#6HnZcb0 zSJ7&H#63-|1$=G+LGKwJN|w_;E8~TWpg`Yh}cIpxt`RH-%1w?rYib>!r>4S zepEbAjpJM{Q4-LBrU}QMP$ofnE(gOO39Pyu3@$ap2T!8hK88-;(j4a@yQOX4*O=7~ zj(?S09ZqIy!Hyb{(^h=2kFukQ(NsbQa-$;Q(u%(4pt}h9_QSpcYZ-#vcW6TIe)(N6 zrN2dYjqn?A=)+g|JukL4 zgSu<~rz?&s@TE5fyAE1X}_Z>BI< zhPdlK9RDluTkO6!4S2ldEmj5&hsLb}q)7f~yssMfXQmq$UWj!pJ!oC_d_Qk}1$wat z_=3Q0JK9cN)fv5o3W@}vCr{9)%snI=M*{|Em+cS-cKgLqxpO^H2v^w4`{Q<`3shPj z-EN2XO5m`lT zdtzhxi(KYbNm4@+ZE^mMy@pUTJzif{0ucCHxZmF#CP$bcW(A=`z3s#AEqiTE%;>H;uY$MVk5#V4Ai=-oF`+A!7RVuMpecRtjUu1KN_9 z4<}xgXJ1zp%?$kvhZ`^`y(|8}r}N3bPpB>wRYC)WK%?krhjk&>W)_tg0GzRF)@t zrQF%IU+7EY5O*HjF;(KG6c}3`#fPascqMJWZl6ljRMijlFJO^W|HCID`VFlHO5y~) zcn8Y*B>OF@G4V7tlP-138q}#MBfy>8quk^j!MgczOObJjKGp9qB|;(_Pa@|-#3(aK zqi=tjV6mI(9>EY?wZr?*63C#R!t>cUo4mS?NGRY>nOzQ&cmkjmCCg)LXBK_+S(+BW zawmO#Airy>k<>l3D&ZJChU-(=!x3P;-0Bb1@@YuLy_75VT89{kxbgDqs)!Aqym+8c z6`lt|N93QjJN3(;3(1IJR?-U0kQQJTB$vFu(m0x%T9l!O0LwfFaq0WEF3TJsIAoE(2*7wz)lz){E4~Svelw)dwCCvCuy;UqND_Cvzo|SUfYZ z(AOxmiUXN&C-z9Z1v2e6D=>G7x4Aca5mR&A0-(gj^)+}w4pw_oxO z4jjrA)*BF_QK1$=Id9B~5q=u$C*Ws2VXJS|RFKA~Sk1TTll-RMcG@W2OIPAL_)Q1K zdM{a>?2#KgT9qxQy|U9n#XJC&{e=o2;}J0rc44S0fty?PGJv1nVqxWPUTux|u8CXx zO!jJOzMme!M0nw1Ekd8L%wG$AESRt2EEd}c2^lRRc=9v#hxsF{-${|8>Gj;!QDIo< z1rW5RFpp})Ji}m49ZH-Ng*^I(is3sFFUA7G4%UV!U)KM;DLCuiNF}f}=mFZ_cHJt@ z-xXu8l*s7la@_bnL$DQ#?Hq0YY-iW1q|HI~b$4=%=u4u)5l~`zUhk5oSG{q1%O3nSHcvZ8nciSJd#emefc_zmVS>R1joQ7wjTs> z6HxoAB}1fbTjGYv2f00Y{%CjtlffE!{c|%%$o6D78NBf&)Uy*l0z_`sb)Ctz+4o`X z9xvaj1Po=H+taS|%alQ>0&MHwukKn<&LMYh&Y17k2L)2h?Sfu@PSn7E){pbM<}B~u zNIMAL2Ku1c_Q2zx7VGZZppNeK3ep)mxpUq%ims=!e;t8QFQRjn6zl%)-#t#E2g0A!f>LYkf`R_|SjjJP zL%h))$-ZbU`rDaA(T?pUmVGPlN~32LaDcm5M5EuBW9_&?-KjV-AdWQTE27H#qHvL? zjx-rj9+MPq(mH0RsG|a=O@HRluaI;d_2z<6XyUCMUC;0Peoi(;KWAe6jkGV*xsRy> zDl~FmyhLqO#Ig!dKi~Lfq%abe@K0-U>QLgsCbNohmD%k)!a$mi8|inNZ&lPr#o7B= zIDQ!G#a)bkZ#<}Zr&Yc>g}h7DdZrf@ZR18sc^{3qbWwOh+QbRfIMEv#+GhS9{;pjh z&y4Y8k}IA7;QCmI7(4qwqyG1>&23hrMm@+}mHa#F%g{c(I6OxM;TaMbJ_*AHk6{>& zrqyaa%VsWlr*+7(F@?O?oq{=io$VC9pw-pgzA>xm#mK~mj!57I5bZu&A-%Ja7a^A4 z(3J}i$W+At#aHGw`(@U0E6YazuIj+)K6eP!@$=6mw)P(z-?56b`+ab1>st{tK|v0! z5}A9}rvW`AA4vFM@JX+i(s^jXQtHhVt?Z5F{p@hyKRFMfQQ#i9bJL#3!7D4}O<|tU zKnKE)nLz)rP`ycDl^v zBjD_(%y4PTEjKJFNu0|Cqy-);Wbk9JuQ)j`aN%r$QvFv7Gwu<1yibLl`jFa7{SAgxw2GHY_-?c9s=K96&2#-CAh2V#!KNt281 zsRcjW@Xqzm_#!Ci8j^%x-ffELa#m%9e^-iyKqdA{8JsAVrW2tv%P>4~a_U)&8%mT!M9_cDvi|fbInt$Li z^o=o%jHk|2FO*PVDxJqwM6=uJIA%%{aq7S#Bned1$~2WGtL4f>SHvrtF?`Vyqv9v4_3b3b6iq1pEK^dp+G_m4bjnpBiC%)#w5*sJ)1Q^T>q zsYs8dXy?-{*BP&V&X_oZZMu$D;bqNN*cfRk)8u01mO&5L_||(oPA{Q2BsvGo0iegR zX17VlA0=%}@8JxGOYch~lto@W@m-||82B*XHNqN7($&xqjE61sey=BqFpB#L&SA}1 zS{AB8H@9m!P+VyIIpUu2qbZ({JMl?m2j%_(`EbPvLi9LQU9a9=f*|*xac%c<@@hzy zO);!rX2r5ysX5MNNz&A3?2L`&^9x!eA=3nSms6Bkix4(nKjQq|Y_4%IkD+&WWzNlG z6e0opdd>!#5!FSp{PO;yqn&5u`R;`U5nQsO+d1`2ht$b!pvXx45PrW4q^JoyHe1K zP*0v`oYl3cWQ@XK9ETz3Vj~pC@447w-?^RT%_OFW!6$`J6MkMehS{|sM}9+8%r3xP zOE% z+YvrQUP0-n!>wHYah7DPBe@CJOhsXRrp2b2H=?Z|ckdv(f%}CqKTI5wo~tE5W8MO& z?`B>{J)XNI2a5kL3=ScYSE*Ai)Vx(V`!iKBSxGP`HC4`q_5@`U>!qr|@+r6+KL{l7 zM=7$0kV;tMA}pSEp}ypozIB*MFnMkZ{n87I}S|O z!hOtLMH0v7rr+NSae*>ri8X|cuKa8{S5{RN85rk5c#@y|5A3oN3s|5U6={PVWPetX>c`#{4;=;8#;Tdv9FUe=*vV0P_D)A9x+@e+#uQ;&J=$@;O0Xp{cc*5P)y z=Jl{SbAqmCE*ewfDj;d#n702q-dipzO-;*R$BT2+}V>g=FX^KcdMxrhV3%IC3UR=re@zWVI4Iq<4Hb_QRIE%_|QFQLl&du z%$~}1=D_bFqj2L%5O}z_;m(1=_Q7G$9>cti$XlTJj0L+@mL*5d^u;5Cl;O%~!W&p8 z;}d)+(-2Nq_=p^TXP^PuP2(e>7!M zbBy@>SZtfnBW)d2Tnn*o9c#~dQhlvzf7h3Ha)tpTL-Y=iQ+^=2Chi}878jRtv;oNQ zQ!cIQ(=y={`N>GusJ1qRnwwM~?`osf7)dLmv`Lp&Q8~S{qM9CKhD{- znR>_gy;9%;;Iznk*1gBxRo)LK*mpu@nJBFi=Yj}VHd!n0BUbmo$A7H0{#&x%?a~0B zt*P%sBH+i>0%YNtVejW!Djifb#Q71*GL@w`BQ&HvNIYqL!8>|~-jwU;F3Z5nLnV@e znyl}pUr=avD?T|Qm$doiygmCK2*dxwl|=>ag2SJPyXEsiuA~Oth(!R$QnT!zp9xUg ze-MDk(r~W~-bntqa#%8045xr5@WYO1D4S^NTS8ceP5=XsyGC*kt%O_UcK*8n%ip`; zdXx&x2`M-OJfC{A*d0?C`La`iSPur(MNss~p5_uaA{SY9SkQ(T@8- zr>+$39Atca9X8n*#EVRh<{0wUf?Yb&_(%!S3`6NUs=yy-Y(g#}@~rII^t~3Su-l!? zI^F$036r7aG6a))bl4Du1pP)F23_4N2`8+>3EJ`Dz*$|a^bO)Ussaq{$7>~E4v^b& z^~UglEF=NsZP*nFR)&ghn&>F16bP0f^eeRQnh}|@6z0yLl~8r#OM-K<1zJZBZYW0H zX)0AZ>j^q2L4e#lHcbEJ67bBfL^9hFK@7XPZ$p}=Z*v(~p00lc{mSGSo!9~lYGq5k z-$_mD|AJBidykjsnc3=6@4vah>_Y-L_(7G%H%~oK@i}!I_nBcqH?9^J&eNO-wBdZo zQvAVS#j=T&n*B%pvD}WENOx3MhqfN63+@7vjOcgN)$|QyVU`MnDd#Ka zcH_WsK0BgH61dkVbrM&{fc^X*hq{-^VKnk;?eM5B#0!i`(6ZC=!5iUL6!+bpm@xV< zrDqcHRT-n$DlaV7fMs>;ZoB-~p^f4ES)OG8b(Y5(eK?rcK-JqGMi~;?ylUyxl}!W5*0f)| zdL@vH(~k}1{`yeWSdLWN;BOZrV0W19ep*sLMxl!P>q2Yn3s15oWsb8Rk%hWAUy$zp zQv(wjL9bE)Uy3{WY7_h4Ze(}I={+ynsqZHba$8IEP}k-t2{f7F<=TqMN4VUy@h-2Q zEua`v2r2#XgP4sE78hAK)(6rp^C;W`iq5qgMMp%y%EFe73=u46?a$gXv3ySU`pQy1 z%%Gf$9zc4~&u&BP1xgo^Z!Sp6aQh}uax;lQk*rL&yRN&cs61@Y4d>VQ+Q9+S>icF| zsIjlJtFw1a&xnE|9Pjj_SU@LcF+BqbDw{@Qp~el}bv<|WWT9tdo6e&l%&z0x6&xi| zm@;F`Vi41tse&4nT+u%A(;Q8#48|-|SYT{&itXko zP$8?tkt3YA>c^};k*F2)0w3R5=D-k9D5#HGcsA0(IH|~MJ79ppDomXeJU+34d?PUW}p?9L?An4!&>rVYo zq1SJ9w#wvA$ssYE&BvU=e4#_aslKAWL;7gx0WoQY6G0K6XiHR74H+VBUaNf6-we() zZc2%;K++^SV-BI!ewRX>7g>Vl1!~uJgz3l+K>d;jQRRZ+IaY1gO2KSXj%RSIx?olr zY8l|Ez?VzEJ6fKoTx350vtM4HZ)6t;qtkOTMZkq=t%f&ZeEfM67>D=hryqHzkd z+FmrzsbS#d#^5p9>mPH;aHdtJn0F9i^6*QQY+;!y{K!`krD1n)fl-tTe9v{gZjz-( zq1~JNRpK#M7fi65OdTMZS->cI6xXU!ZnWR zd%q51zlPdHeq62M66%o6{m;MQJFf8C1dTyQ`dRN$ghGB`QhulcJ zgBovm-9i4g@Uk^)ksyDZ<66|q-w%_0dxf#wq=lmB1z810=i&Mc^-B_Sx_8a<8>eC z!@(7nK2e-pc2jBLI?$AbF(78c5mCs2NqdWmVsreZjFR>7>3yV~HuRACT|<~|DmS4q zlu-q4#J4L3a0nA=3H9SW`?WTz`vYiFf)J0P*Qb7jc~*6yF`Iev{WQZMHU{in(W`Jv zhG?$cWGz}3*5}!?@4jJ#+VV^SZdWjs;8n;&fzn9@W}|9SgtgGEvP`sBPrNaCprW@=JL~e zrlUQAUP76jBS{dT;peUybirUy+{+zdOW z?wQ##ALogpDkgI2cEGTpa^7=Z9^WnL5vBCt*Gp3BP{k*#lCJ|v1um3D_KU-OIHFS$ zWj(t*q(ZW{K~m!r@f4jdqw-{$C@TjOWcC~NMycnCe$f@z$pKo1-2`7mz^L(22nDLr zs)R(H!(V;yaU#bFL0sEdr&;19LkRX9HDEXouC0hcCdf zJIw8vzul^8x!;^LBYp%^x%*fXampNw@$F4D5J1^JVb7APU`3u(Nfvl}zusN>lg|j)iyjjY=0>a+gGd z=s#+=0Di(LIWxUzza?I%(!VFDza@6oA?10tGV5k>jS63>HOBh4{8w*cM>>sN{lCg7 z5PY}lVwG33FGHGqfU^Bg5LS>fN|J`r&JA!8suf7Ww~bZkMu(_A@P#enn-=hQ$8jq}z=`h-J94o% z;Y0Tr$fE`HN$M|P+Pnm@E&jaBi)Gh#m@@1?7735#=D+>nvh6ED53qfyY33zVPv}FZ z&o=KHrR%{=gBdO?^G3sga1mZDw3EA8*B+f4Qa6(S+}h{L@rTaq%}4xCr>O7=yf^l_ zBa}Y&xl$c9(IY%9eLia4ZLxSy9?a|%&{NXz*{!XkdP|7csMD$D16r;q>tCO8=M=JQ%^?3lJVs^U>t2|pj2Zl@B zVVR8IYm_m<$&!)WTAP>5L4AG$UZ}c0o0m+8$I_Q_?#}_L3@UkkdAB`-2Kwu)n*Mwa z_7uhYOTST$soVPc^zD3$mH9O)F7^Qkxb@|>z?^g<3Yud-5A8gAI9`#mCaYti`*AK! z-yQG7MgN9gWs)37f#?O&ODLrH`t61qwih0Y>xKB8FRE(V_*M~~rZN^dVU?PW81!8H zL)mw`;$G%tm4SFl-z_COr!@i%x_{xVnbp49G)PNne0npzXt(cjinBH$fJ zq2LJ9Iue;_yKmn&!gtHZWAL%KTWvDfH?FX*+%*2;6Pk;uddD|AEi;~;`F@g8?C?5dH}m>YZ`gGX;)_5;Wx(oQEAHZLanBTXo> ze@rO`R`%cp0Sf2>UI|s9zDCY5hH^YZXTpcZQ7#i{kSB0xQ@Cd}JCX`eQbbZxDW~b- zyk)$T>XrKX8g{Eg$>cfyl>NJ1PSR+E4mMV6F|A-~UvJi0$d5Z7EXfs`mM|kMwc-fd z=QFo4rcrH@casMbG7|Ua^qP6W!<+|b76I&W;A{^Fe2Ss;I7MqN0yyqG4pu0p9o zKZ#F8Q@I2*m@lcxA^AmWThCSGe3L$MswgTT8t+752-3&*yd)C>R;okhw|!I*In>&8 zRb}}~f(ro1vTlnNM}?hUKE~_(6b(~1F^ZmHp5ze5fm?7Q;8pXRGoBToauobM^A6tY z0()L(LLT(7Z<{Qe9|d2h)6D$&a0FiYFi$$|{N$A9Vx7?68~ehyF~(7R+5#0N&sRNP znT&AcR+OUOA5R1*i%(jUQ^j<7=1kf46?Dk2i)O5RaxUJ*53w9>H= z&}TRWj;CDPrQo`y8pcfW#Dx;rg@;Ig>PA@ z(N1!2%|2;+zkGiN6KxEK7mn?7uU^_U%AA#IKAIrTi=!ayx{H(_I~wXGqf8OzlT6+* z0nyz5zb{!o(&Bt+5->NPcscNMLD;t2qTTnucl^B6xDI_=#9Wpya467UXK>=>kFlfR zb?RO6ulW|C>vZyJ2X|5VtG`dd^-v%)#f;8WGn0-)ysGA_&mUY=gw`LglwfqeCDofY z&9@~OO?t-uw_AWO!xc-sg8UmCG?P@wCFvAHH}!nm5X>N*G<=br60qPv&K7~N!QBPc z(d#?RKf%%@pC(VWV?#*s)B?@g5^#|B*o72>;_v@W++i+R_zY=2t*1rSmh*?sv;aJd zUwwK64?9njhxwKquY7ItH>&O}n-60dW0}~|ey^E;9;b{~~Pb~Z81?*5szgHD<&JfjB8l3282ziq8UNaU+#OQOAJ9HDJ;izok!n{Tr!tztgTPJ45}Y?I`L}@ zZxEs%H`f32pFzK-8H(UL>9yWMCiv%Kcd%>61*G429X3`J3{Kfg@-UCPqh!baY=3HZ zw8g%#=W0f+oEqz~VrPC|c4dT?dHRT1J<3v}dMg@)un!X7AJ! znZ-ViJdk|*>VcehlpbqznfNUt_SHUFWAT9;HeXub?k9Wl)tH)veu$HM8%jlHk->L0 z{~gspATM_y6F=9e?vuT=zJO%v%OU$Dj0@*A$QC7#Z3O)z{l{cb&~SFqhAjuNCy~wj ziiHqW%6Ee_+sw?CYZtq|&|V9nM6?{wzi~_q(w5|6^uC8#zLxvzcCQd-^i>Gacm$bx z83D>EG=SEsKBPKUO;zMtC&<&!iU*fj$GB4X25MTq|3B;HF<8{0WXHEC-}nh`c$zrC z?2?^Qmda~OyL5(uoX%T$aW3Ep?NZTM3}Fm4+n)Q|f;831f8>w&h5S!(Znq@hd};`# zO#aMVs=!c8Hk^ijf80s&^lxDaS5b^s_$KL66y@lrLl<*xi`ivyt__! zFcVQ7eKhvosdci>{bQJ0IvK38qgE&-ObvdPfk~8vjd*Ht2|5qWTzT%q`!U14CfCZE0pX6fzQg+QPCT9Mu~oQcWfYw3!|BQBq#!WVktiZFPYUsg6I3L^ z=6kvU_Dw$gA2gi>S5*DihE+g71f--H5h*F@?gr^@q>&PkjzLnoLApcfjsb>7Lb|)V zYX+G3&-1SJeu1-Q%{piQV&D6^l1oPvvlT2pny>;@4?lgOj^`_~3}3`9p?RC-O%z}D z6)Npn5-#D*=pDi*jsEOKyTATGgiJT%dVjJ^7@2Cgx>L(xq2+XU9bhw_h&J4G;o)PR z(lhg9d`+SU77T;L{j+#zELq*99teYk1H8n(?B4GZ`j;?+YPZUMKRgtV9@<{>D}Qp>|kPwulAp{3wEnXL}fH&wh&4`{j!J?X^(&?emT91N(&H64&QuLCjyfMeN1;etheD z(HPm1c{>yN<~m*II1e^>T{inY%BUDP^ZYCf*eHJS!CZ;Nx=&n5U*nQ@lfzdaypyzR z_1?=*&Wvzm_u>qFK55Aez93HaWn0@p!~VfE_(@Wr&V`5;y{=t#BUR7z$TagI$xH+A zHNX&Ar9)eMH1FSbC+V{qJv>m0Q#VV^?nZE`5x(yd*6)x3Ri3 z8xkFX)Eu!&Yxid>9>Y{X2FJzA$OXC?hX}Eg35La=UuH)Gu7~AXNUlmuSPPD4%1i?9 zk8C9XP)3)%1>}Sn^0j9NJu{(4H_pNCCjPHKTFB-ZJn@HO!!tFfNqo&cznQ^UL4W6! zW)*9?oLEZ!wkoEIw&T&f-j7zm0;*hzA9}AfJRT{<3c5ZUq?Ukbgi2pGJFdM4ea6x8 zczih9*#DPskj56dZh5XLRL2osE(o9f8Y>qQfa016dcKjK#M37o+fTH3A69k@kN=n7 z;4yk2m8S+orKjcidju)X#`}E*oj&4tBzAIKBB^ICrNXV z)fZ~EAzM6{4Kamh{#hF@Gp8`k13P=wi1sa=mFwe;CY^#L^#NMXj>T!hbq~jT_u@p( zgS{^|)DpLrrL-IU~O$vk9**iMGG|8hp?nym{3ML7}_-jl=D$u_{7VK$t*} znx_env&TRysVw=W_u>)fa|IP9Bg}^!}lxpTOiag$$IEQb%-Xw1~r}m^w7NVVnQv{R1WX5oH&8 zh=+EHk5-5BDy0`Xwz$V)Nj-qV56%GAky93T|_CmG;pdy=Hxj+t!V1>fQ6) zccstRv1I3|;ryn+wE*PoMH{6HRcUlEI~eu~k`9K+lHUDiK*L5}u$!JPH#r>S2KxS? z6c6lhh>i|K+#ZNqntNfsgtlxNpBhn5ye?_DmN!N{zYXTnEwRMAU>XuPvxbe7B9v+x zLvm%ilSfGlF}U}4ON;9{CgmkqJo^pPt9sN&BO~3Snv>Ks-jdsLqsQyjEk=e3_|^Nh zoS)}-Axu$8?#PhE39+Yazr9qw)ouEVb~D(vp7U0(tn;J$Okx_IGdUnPuEiP4=Cxr4N&dQ_PIAW zf7n2pZjPpdFPbZMw432)5|6n2#uv_AKCK`?y4MuRwi&iq3F~7hr=mzYl%=sUM>8p%1u1QiA z9J`WL<_%<*-wB(rS&Ds;Gc`SRBBDPMBFzD`Y|Q>5{X&Gbniy(a1v3ef2EOg zsdn$ZJN`@f{CjpNv5II_cJV}m%8d8^xAR#3^>1Z4z7*$_`cabqqE}HU!t>EGs|jm% zB&q~v7}Z%=qpSZy-r0Fg2t7GM_KjTVVcf|Y&1DIw*I{l(yrDh^pJiW5h#6TJ7LmA8 zfOA=@24t6kLhv96OhD^Jzy&d6Kg)#y=HBa9b;@G?y933IDeSb(%OT*}p9Jp2E+|jo zZ=1g~do*ki-+m37;UL92HmQT53Fzgx-T8-jmjvk3$9q?0E{ym<1N`gjbC1Xh6cL22 zOn66zKh!7+iFU3ku1Ek?w>8huPy{XfY$Rp5 zUK$DLsGG=rU^3Y|2x{wov!Bn6MkYZLYWHS@YDNM1hlQP2UnaiOOA`+sQOD9x^bN%# z#XsNfj%|X>g=ad8Sr`r zYy|+4=#iD2V=V{rsQgQT{2&mrGJH)+{v%y` zMFmX{GWe*B2PHI@`+$y?(p^G%t!kSrwSSO&wNA(i4n3P;o08EiyLHX@s2Jrr5}^>x zA?I=+h;PIBiF=-oJj?q1vxoRo4y=&2VAvieYY!T`Hoy7kp)!NN`XR$RA7;Im+z&Ip zPL>KqyOHb{qQYKv%ixae%=y+Z>V20 z@O<;T=xlkq_fnk<>isW-&u20Db9$@4R~`_sEA|A_s!Uto6*KJ|apn3Vtc!hQGB zsI}OWS1@CFXn37f>$get+R+7O#Xa3(S&LsUwH^RjTgdUAj{NH07HzRp8Y=+yQ`U6G z)s;fzSxj4j10e43Cy3Wdvx}tZ$b;4yF>T%~V6@r(MsBLbH(^=)C>g24BJR1m)(uMm zFou+;lQ7}vO=$&YX^gWiQcHJgabWT~?{=YtJ-tt2yKYa^MFoD>oznqvJ=*8pwLK0B zJquuD00hvQaPH*wZ%z;7YyxU;s#pkWQRxQ#yWRK7_{Mce_xht2um=Oq?a0@p=^q$# zO%x}h8CYf1GT;1dS&$J|AP83FlDB!kbv!XH^oV7&-t;4(n31$a9pUNd+c4T@Y2r4W8dSGy@csYhuYAmRyNq;FqjaTSSDFR7$dvo+2 zuFHZWFOqSyR*&Z@H2C$tzW+x5ve4i{zeNGK_UPe^X;&?`_Fm*}1~o?^VT(pH_B7Ym zb(*7_@Kpzqf86Xcwaq&lU3Q_o-o8qSzHYWij)J&6mG+aSFSd2u`*Y z8bsa{;jfi3;oLpaE@PeZ=!2ZAE)&G3FMf{t39d^Gk)Fl-*aIo~4dBbLBZg3BoKx** z__N{<1`(DfSjvdV=r~ZVZed*IMy=^RGD{xdku6ZVx>Q|TovbG`CppVDFODFa9r&V{ zb-oJe0!FQnKH5@?9%P%IAL>?Vqb@xNHay2%UZRL%*3G{EzFsN%l|E7VT6eDJH>biR zyC%KrpD_8ltcET4zx2Ylfj}vldhV_bZOL}&lG^~ z3Bp?__#WU`z{|SMi7{Z?f;{O0Zx>1K$x{`iGLt5;dW#L$Qk?~?uwDbY#CP*HaMQfz z$1={rSJcK~0>=@aa^!WLpB}p-cvQdo!>+zoW}a@Vir#6LILQT=iyFuKdjC!D+U@e3 z)hM4SH=EpQt`c7Tytyw!u#%*cfU{$O*lbS5KPEaeD@QtLylfUPbYF%%i)zOs%hF~> zk!pvi=n#UDH&anSDVi1S^Gmo<{5A02ez7X`UR(;5G~2mXb2ZO}F^;#?y*Bbsp+J?+ z6`Y~7I7EyC`yfG! ziF7ZII0J1ow_tLRAmPlVkhJ>~TVk~fFci(L1tdbo^;1F~U9^COX-!ppE$92TV&?6EZtC61!XJ)h8ZfnU=iJ-UU)B0Z zW?|7kRVd2woeOX}Puun`ajW9$=}0MtWVaDnUUP<;Cu$`dJMT$qW(D1{fACj8Dq%4t zce(vkI04h<9~Ik2p&I9r{+JZEAH|e#?GxeHAM#hjjv9V~l%7~3!{e!5+^r#OZ|zst ztN-tM6e507fJ~hOM;+&NeKh-ZF=2Wf(ohg7)&zU!8b(Wvxj`ttLz0EV(aNNv9^Nl~ zW2u%ccDMm!PV{df8hhUCd`QGHL7vO&kd`HCi^sswRV0-YH`l@OCCFx9nR`vc>?!l5 zzaeqUtddf7>8f8|XGl5vMy;eYMCc{X)!w|s(qAKIS1nhX`*kvOe>5~}`qtay<`sF6 zrE9@NhHY*b3B58Q9yqk8V5#9i;5-+aUIRo6K_0aI2O?(>=OPEunGS{yOP=4(s1lu( z=zoFnT=Fee(0bN(Z28_Fp40rg4HDk|OZRDV`FTZhItUKg0E9F|w5fg4hm1AalZ84& zH(BEbL5$CQeIRQCi>`HKArD!NN`8M6XwTohuB_^QH`{1iYL7XFcsjd4Kzjit4tRlU z0%N-CT1Y=$$MU&Op#PREB-X_rEyO6oyw*9^Z}W#D*&DINPT2M6IUfV0&(C8Z-0aI@2IujGd87<>D8D<&t0}eA+{78R+@1c7}#gP14%}cdQIC%=J6k}&r+x4j3Ap;vfy<)F zr-wvlq5Yj}mO7%B6{}@pca6fwlxFTO?2i==Gy7yIN3LE=LC+UexHOYhP>6zR-7`1! z>M6%=fPuw74n2pO+>_r; z3@0;!#9><)xbMczK5bCHeo*aGvKt|l(FT|W?CC(pH#QQT0ltS{Qg+_rTDe~zX1k9+ zFM0nuFz|f;UlyYhX=5g7BIl2zyyM7B~^F2hP zta>n`#|r^FIQOVj72n;I{`Tb(M}QIOSI&*xMun6wR`tNwBy*Ii3L+R;7Fk>ucl@8~ zJU|Tkf)c=|vzE4+0u}E(kcle1i<0mAFJUz6oG>9?8Z;!%7oyR0Hg{l%qaZom`a6f} zG%$X}0%Fe$ITvubg?YXJhBf*59&4eHHcvoVW{B(z8hkBq3gU0aaYqfhzL@w za6tm^$6riVSUl_~m%U>P;P@D&poFckgwf3O66*ixxUun;_;lX3CAhH5mXU$>U4Hp@ zujq0GWh6LnsR6vv2KM7HPWTjv1DPE>9UDIv1ZU)sL*Om{337p|XK|c8PeIS&YXj-V z$706|6V$G|1aUrL=CEQd)%n)7v-7SKRY>g%dEIh-Nq4&(&{1g?Mp4kzf>2Qiogp&9tx9IWKFn|WPJ}V)u--F$Qu;{U`_kuM@aD?RYm_1x2?ja zTf|)0;!JL$U6QnGWMMUSFOl$f)<=5MU^nFSBYZJA{$x#4DyoIz9CMS$ zgHYtsGXp$)FHakB?Q zUx^p4$46nLmu0yHG<|tN1!u+DG(1q25MV`cdW*wTb@0x*%RdQZ=3E8n-pL5G&M~SP zis$rPtWI7op5{U>K5KJ-EOTdHAxOv}jW4d;^o=KGAeuxa%S&+QV^S2bY^UiSzc(}V z)x#LIUf)Z9z02;FEkBangc3%GjhI4>z*fy7k8dn{IF>-IfE_Bw}G5|9fnDF^t2g& zo7%tRUdGOpCk4d0`bd1XR*E*{Hwc@>$ zcnOrdGKV{V@jUDHjpPFr$3jcsDw{4(YyeSU@QTnI(ON;a=1XRw8O>kFFYuC}CK8fy zCwO$a+U!-;3DdK##QtQ~7N7*Q5|}otVf23HM9mi-7CUp1w{{J$6Q19ILewY_fTyRE zntQ}^z=NDZ2Zu)w%M2KDCI$Iqy)D$Pl>)5x&uidw0O#Zpb#@}fKU zhX(DS6@8ui!_H0DhhF_oz0}yi3#FLOD(O$x{K=DE8&>~CbK3TN=eJM8+jEp_#~p_r z{b~QCX6^C|-KyN#plofn2FncXx51=fK*3CzS~7-|hSB2Uzm)=MKM2fyBEth~?|8O~ zPIf@1Ei2Kw}ir2cV;A$xN?AMFWu*LGhi zO-7qPbT}UC8%`fPT+f{w>Vg8(NxSLs7@a%`4T zl`NY8`Wm~b-sP~FzoB+JAN0}9P|6Br4i6F*?F5}s{Q1GCJ(jf!WWc}2)QrU6CQmH% zj^N(Ai;IvIYNj4);W`2+{0WJuGP!rU-FklE&vqb!%WETW#B^ExV$`i88jCBEd zI_C>+y$VB>9@Qnu+*BktH~r7F!ZQPAr}gyr zrQzso>1Kf$o(jkbY9@M|Tlt>mZ-^Jnz7!}9KRRFIU9{)yjWiX#ai=e)Qn``txi?Ie z*OQUWuy%4zG`~7@H7|0UBkSG1M5^@gOjm%Bo`&Uo$>`F4Y2@{3FS}7`Oz3bQJrn$I z;ILkz+lwFmNfU7YQ}J&>%ty;jPL?paRUFT5WUWFJRY0XSDGk!+kXFkOH+4F=MqnRk z&3stk9r`-s>e zR3%ejDpmK{8Z&B7$EnHAqh~a#$IhSbY(oEz{r7{gMDWqqZMGC;*rg;bO5W!S*~Z~e zGSE=&y)7zUEq8ojOdNE&oa$)!YP@cF3eh0^xM*4YYf*2Wm}M=RL(F9-@OEnGSJH_C z_GJV208|7xTEw$5@W?&ZX{ z8H%;t{16_@k8P-F>q%m_kPov!)f1d(e-i94@{=eMh}(jxdn`6W85)Dff%`hrSMU7CA4A^x%#D!n<*H=zaVm!mCzn23Pf`JZu0TQnu>Mq-tC*{8KK^@MoR{J-rosx(^}S|pOQv1Cw5g)CUhR51uL>lc zeiV|0fm%k1VJ9W#TLwQ~Jggm|cCnDg+%NXu4+|o&gVi?UNo3|I=5~8a8$q96{La*5 zu=xE$vsLac)21%Ofz@5SI8$F|u~Cn%!?qQ%83OtIqiMlZ=X2fdFk@b;r{P2auul$G z<-ybsxopUYKiLkNpXZu4VbwBhaX#a7Ts}l)+^y$%?g~b2%VsqQcSq54<+;tfzfH0ARPK`3Knbnw< zszv2nYeSD5*N?7ocG$+0tklWzo_COV52?!{nx9?@_iWY+{vrLoq8~8AsR7Ty*f6gB4VV|Rg`;@}x!lO3?QJEK4w;oStq02`y`IXU(WiF1 z6}N>2RORxUC3gc8QpmxGcH>vWH6pay#kLbCgqQZLx3MSBfb-dH)qMpy#4>Pn$VCMR zBo<)GixI6CY8B=}ZDFFWm+W-(d$8O_j@R#PZiu*p<@l{yFXY$c9@|ddWfWx({|6m< zwOjsGQ3JxE59T4Kc~23pk2l%5^n{nGrv~#eyY3*u&W!S{{}!=kA_SXT#PUg+X{1dn zlSygg`V7ZnLx24HJ8pmE;{_55jdtD!C-`#+Iyd9WdD+c8%L@MP06UoaDK~$pX0-bt z4Mh}-Zr-fPn`koMIT>wo=TGK6Ggs1CE@d&vI;!n0c>#}kdwzUH}PekZ)BGLod6uL zhaXd#P@r&6dVjCp$1EZk6f$&S0z^jrw{qK_KRno0{+4#W*tmt3Z6J>>0P8Y_2X`BM zlCsgAIoFr)qjVq`@dPo$yO{vXRoY$iZUYeH8^DL2{=n}H>Mt^aO4J*b@mOVtsiGL) z><=b!43#j(#To0%;%O*#n8QUcJ3p~$2oSGsT`m=bHT$?Q69{=6TY(@y;={VXXi9B{ zTRx826V6k;ldR!e9R4@RfL+A*Q9MrE!#in%39D~!VSJnS{1f&}T(Se3Iq__@*brzmVazXGLLGdGz;Ke4XvNY zQN&n^6?MO7JuA-J<>KNRTbpdZb8|2D#36ZW+@}~Yo6$XSw$jX9u-FONuys~@|6*lv zXTL(GZky1M<5PhiecQ8Mr(YKxcJE;Lg2h!J)ffGge&6|NZ*mE_#(#-w2@rEZWuxAv zb|V;D%-8*iItFgA^fBGj#c?($Pn}|dR(LH zj*Z^uS05(OAFNEchhk@|H_+&f`{2VagdLK`p61f=Dp^>wtijP4-*7?;4`a9|Ql7O3 zw{ba?QtzO>#O3i4p`3T&ftYU&#yJn>wL>*G-jV88Sj#qOKR4t5I>aZnY#|#BvZXgR ztysHMuqH=xPn>VB2cD?7H>4*1=A4A0HQbQrsN_dJWz4mva1MU?LV%8K*4ABUlZKux zORrF)SrWA(Juh*3%rGcl@E+r?qui4dOrP#tVl|*8C8had#4i_3>`F>^7}3u`gBZ4eUT26yAqlSt`?cG*VpVSRR`ojHQW*|2G-zg4&) z9ZRPC^uR?ZS{@y>uk06ur_>DsPi;!ac0*N3E!kj_qdJFmYrwL$jFFNzR7-yfY!#dx*nf` z!B@!b0p;CehmwjWwNe(^%mY0zJVMaphLWuSQ0+G#%`qRn@usxsS5JG<4E7MHtT7p2 z9ob$)5m<>s!s=KZRa>mz@}={NtH*D$T z=ReXte$Ek_Tcst3$F^x3pTxJCRwizwq+(s59<+=1b-F*tM~;+eo*Qu)hwW|dHHB^} znP9AMtuc?{+37^BW;R#&Hl-+M@o>Chs4=cnsxF>yCvUJ(O1#TR`EX~9kisZl)Fgka z^rpFMG3*L}pMk6dwhlbruOvFB?vps6_5B$57e;ED-=;HrIaa2RT}cpw)X?PfIQ;g@ zZ?o&^AMe~q>b3`SP#i&LZF!ycv#_3*TC3ZJm-210QU*QC81ZNTmIg)57oP?l=JXB$ zftyR21!8&mknWa^7~(Vetfq`1G^}#U?xA=vwJZ>t#ix(M_eXU9#mMiSIN`xiB!aI0 zM%;tX9znvU*`IQz5885ejA8|;=a0>tyRMU<#*k}xvAwLoRjH%Rc5t7Asuo^+3I|i- z<|eD6pOH_TgcoM(Ba`6#!5X(zB|~uZ5uA3EeUFqybL`SK-Oa>6frR-zed}P}UA@#p zRB7cwW>=2iy`W*lpI01=EKFBbgs_Px{C1Pe37oYYueH6AC^M;7m-Znk;2=DyR>QkL zr3x~Ffv7I(%JR{I-c{XWH;j?}r323iRTOi|5n2=WkC)x}4(YvN(V0m;P4kpZ&JHLP zz+lb;7)<;z-NZNYgl)75L-QpXz*upzxz=vbWJ^lcSf{EEFif(89A18@mwH7!&rxoV zrO&2xxSqLFdjkfdYSpdkuiucz*A2Vb^d*~>84ai87f~6HmWZo$d`N@GL=#f0v;U4& zD}QQmjC>?dvdwDEfP2nU`z=nz(#xT~WsS>7Zkz8R{sh!puLiumk>3R> zi(ZsXH0=vEip!j1Rk?R*?DIMGZ+Qt+CNqn_|4{jlg&nLuui?dXKQxM%vxNS54c9`O z$AUDLB(tOP5;ZK#kP~JoM}%D33;#dUzs_6GfPXNc;iOJZKM?N3tvUS=8ubU9;ThUd zBI}1h2GmM2px@U2>>d)D(c-5PTEq?W*$#nvl$M zDX@=Y>r#)ICauy{Syziy22g5R^_MQ@yYUpTOedS7do zu`3FrMdZoR6XhrphA@84wR85!1gsiMwp?(syz=x|`1|*p%dpcp%8FCI*79$!tE8Tl zr?Ejt*z|;^+@$3PS>f>avnaWIlpkKGft}X0xBrT0SI#a5eFbBcF)#4^m`#7T^mEon z<%I41bDPN^j(@S!_{zjulEIby>%M)Sf9ti7&#(txDMJUyB-~aHJCN&7)iaSft5q$l zr8)n=?Lcgeiscj z=qvPV%#3;4?T8(bg#r`56^yk8Jr08}#^=tUGd`~@&$7IH^W8O<)~VIM#svo({!g)< zVrmI*I#T?fklWR5#}v^=JwqJlrI&MNonLj%#*u?6IU)u=;u+TRcNFANL#~YmZ*$&R z5Bd<+awGHwHw_O$;~0PC6I0m%TGU!;gdS3oSJ5Ei=3OK4Lf1?y!pUauKquO$f@iehj__REig3)w8 z-UvW?tp=81gsls6&xB^THcP2<>y7;OxgpK0$Aze$d=NOBEC})Z4N0V*M2A#)-V{AE z_(R5o+2VM#8Q_`NK$PcAk+WBQIyL9H2H4qX0b}sk>;4a<4`IVQq}x~Sal!g zMOcI#m%p1uoplLd<8A{tMXay?fGp^vGOY*KAWVaCp?8cq-=Z!Loq8Yo;)SF@=R+B9 za_{S^p}zh6Zd2SQY0BKjIBnj3dd^Gjxe)w;)(73{i?1tqk5l!8JtI3-ZNJ);-1i-> z#=jeMrcB3w-0sZ|7#A!YI;yv`EAFzNyye$cnkdAoqwnswo{<`O-9s{Cg+V)HGVSVC z<2hiR4tmYZ{u8&}|3QhlTV2A*vuG&JeM01^()_IlYS_~Qb0lX(Hp)k`BNeFVc8*TAIXRuhm_Hp4TNi)xGUqsW%g zJKUziPs+p`_G!FhdfpT*Cb5ltTv|ghG!F2bKfKs2dUtd2TB@RR`*I-9Gh_U=Vxg{` zt4z8B2S5PtpJO~~5t%f}p^R^xCM%xD zD-2$7kDh-X!+8gav6KJBu8_^iXGHd%WLDH9KK-7mXKx#$X4X@^q}SDM&FDMZEQ%Mm za31}oChBs$XQP^D&tCDC)$q_XV2}6e9}i^hZ=%f?NNh^*-3c$x04=Cf9C_-lCpRB< zXRRaqnr73IU@G5L8k({dV_2G(}X29f5)3kak%H!ui4O*igr%uIU z!8R#a$MEM?*onjyFV~{tB-RsLr zZi}AK<1EbDwM!rX0c3-^7x{F%H1bv9*AZm!d&LC+yQxjag;>UzCWI-5dEl}PZ#&L* zktX9niiQZp&Cf0~dCoo7-;lJQIKdSwF)=ZRiuyjkGWe=IV3sw&`6(GQX{2;b&v%q! z+$i{h1QsZIcT#rv2%6uxg(=sFe|pNZgKOxW`Mvi=mOYrcFQzS=Js^}OkU4_LzLh@U z0D^ZNxcn&{nnBiC0aKpZ_@+XL>Wu8P?Lwc7xtJw1`8Ys_T0P?VD%O{c@t&R*Y~89- zE`OZ4-+e5bQ4I&=KW8>i>Dd410^Na77kqYQrmz371(pQOFF2I(zrr!!%*%aOC%cyv z@oa48!y$wDm36w=SCGwbwMpY`+Rb$|QiXoLR_)-Yk<^d$nP!Xw5|U@C8^str`h_33 zs$y~Psh|~2AWBJXG3j3+_em8rdZ*_DcYo#1JcnNOG7hGua3f9+>AGEiiz*qCd+rhb z6k>$+L5Ygn^CwHolaylVew5Tr z_WT>Sj2Cl$MU{g8;ZX85;bnvyI)t4p_QzF2vlx4_c|{6(%0r#MqRO8}MUV;5hB)6P_BiN?_P@m~o8;5|+1Fmln1n*ntdCAJ+kf+O@3tcYnPisSWSTnzT8P8jD zxMcuJS8EHi&uYSmE+(W`K4#Czr#k^y*9M^-6xQ=ZGFGx*E4y|WDDhT0ym&p1wd0z8 zpJ7q{J+VZ4Z^X~Qs7ow>*17~3qz0%Gy+Z!c83W8?n$eAF{Ht86)vedCP^;7Jwl!pw z0l`ZBdnjIzdK-pVpAD*7?29_j1SOie1Hy3La^wmP&p+Xly|cdj)b^@lY#*q4gJ02a z;YZx^AfKI(+v%KkQoFp*6*UL!oDYCb-$+sguj3b9sJT6byCpADF*gdg`!5j5~!`V?;FN( z^7fogTXbp+GUF`!>Zfm4{x@6RV5AUjm(2O4Mhl3{JH<5HIAV5^*U-2`= ziJcRd@&R(FyH#h3K}A~Ii#}hvVHqBU<+jVda&LXt56$?2Qiq@W>KN2R^WO|fz1?%Y zIG~v;jpXiDWeP5Om8JfRWI*pWU`DiMwcJ>2E!4JTS4&Yo35nc$O;$){ zzgjG9w>+N3_t#|KbK6CD&rX(_v#Z6k&>#Z`j=cCob1oh)a^^cM)nHqb`^UCi-%J4d zbZGz8hNE(DlDRDNmUJLt*6H!-2!Af^S=W`lWms*D}>Ku=^cO z=-AE@A3pcL_HWxVDHghh2W0+J%s?67vG-pD+1lUbE2)JR7qf*W z2o2$Nn4c_-SNuW#4w#83?8%EnvTIX;y4)6yFvr6`xeDmHMIuFR0WD<~x18r+JT<#F z*Zh5$qE z%=;HbWxq2tg#C%>(P;-a?z_Lri{rzj-dq>@7To3F1BW)_Yooxk=m?A(<{Ix0wFJCa zW2}m$QOEK%p0B}`0JWAiHwP&KDN3w&SxN-g&B54UIz5F}yk%u}r)_2|E_&q=PFn4V z!0sWgZs(YMV)?)46CzT-POfPyafb2joNwhfM^!fy-pt!-g)(uAAqMv1Mjy>p`bga7 zw5qc`Yt44$0>qVs6Rel=MPQd+$BT+mg`EMe8ze;UUf|P`4~xff6(w6AN~n7$$@7ow zh8sSbDcnMiQ_4SwGQCRhmNKi2@Z2~^34XVBJzLu581;&#IOn<_BXY+jXBd?#_!90f zs9!zF&M%9l@(z7(0dO$*Yg(3%6Y;vFc)XrG&@I(NMPAcrJ$`@P4Umx|zd zzLCZU7!$*n#ln`wioC)XF~=8>>BkoNNtrtq^LpS;K;J;4bdz+9V?ENj^F2#&O@zM4 zhhz?`+xQXpSI`Kz2OhaFdPl~RSKB@>5%k68buyN>lI|<>A$H$!!$l$fXE=-%q1NP3 zCiHD=Yw9R%Z+u#zs5gWH(l`#n+UiSJI}FQ@(&{I@Q@*m~MEenLLRmouw!~uHxaYrv z-!muz-BJsa!HTHllg~g#t~~We2H6B~B+Zih2G&meD1VaZ{IVt&hRgMKR7y9UNPO7~ zw+id3?ZAY=)&|UnvBy)|HXumGHR5MY3JLZ`x_(Q2z5;+P66t=NH9@Svt0hyN$trv4 zMEx3{r?vErHVu2uHa0|0oB0>nj1@a9_L9n)Bq}AFm^Ou-ou1(rYZ8W>MwJiFK3kF- zTtf8fgu{KMeSr2;Ci`#%W>*?c9YJ50Te_*I{aaW-4X8-P z5Bkh1m=ZLR3^ZHoyE}D_?ZkiTfMxnY6U+WKf39F`&Eg-npc-ubO%WaB^AEFS;A{e@>s2ZnS ztx_PD|DGL;h^f3seLho(rtstX-w%;=x0r97C;@-I@QIHl`!%EHRwaY{^5Ej2tAAUJzdr*)RO%EmxWPmsw~T=BcqU)ch)(+ zKUcxm<3WvrGaHvFbGPOgpN7_GoF(IlV#kxhUMi|A{ovca9XeZ84ltWyDCDd+%tqrp z6;5tf+_A3s7oY5&SQ>qS(eV#orC#@-+klNT~RM*#wNkw$G~g;;hl3+B!c+Tz$%6C z)e{erQg5#rL0;8M8v%*e(83hp?m{?4B4S zkn~*Or$@6RIHSExC|(1{{kG^Gm3!^!Y6}<~hlS?-ND>lPeu%9}L@O^5#7m5#2Q(7U=ugv3?sg z*lcNjW@I&|%JF=Od~S!*oMmL#@9Z!6j^2tN|D;<>l&2{GtUVTFC%Nq)h1?$8!K+%d z#A9quC(!g1%`mX{2yS4zj zrIRrPoRt2|B=P*_57c75{*31X!oU6CUM2|7CJlN+M51Q)D6 z868DJ$iN>*&i3sQS&|f8dVVzRwZ*2sYWyIu23m&$VsHD4S1oUW*1n#%XWo0~3 zql?@bc0?H>7Tyy4adOLv8{J=*Y0M%Jv^99hQ1`^+kGmVX-{^WsaG%JuGKyOz#r2VL z)oRKC<<`)8i3?N|BX{0;r~CNhkL_71{Kp@ELRMA3y82SKbnbkH#r8r8@~hkzqA0*R zrl0EKvNIL`4?g(7{Y$&?sgETT8y|3cIArJ$S-?MT$`s(Tb~Ew9{0D#fdz4($&Y3e? z?J~z|x*<{|eIQE>`a5jIaJOf)kwPSkD7;DZo#&1QPvI8oah^M8s9AY%{q)mMM+%kX z_+z(cSZc?C?i@sH>qhrhM)&pQ8-@Kf>kT(lMbOT)KO$}9cD zI*R3N@7}#_jd(|!5czFx3;&<+_a_8^4LrZVPg-mI4Ovl&xG7mMWF*ep`}>OqlFjA; zepKEAPV>Lfj*pI%^)EgD;gIzDXEOvQQA)pmG%IrW{>|(9ZwpnH@89(Lv$gfdbKaJZ z)`k_y!}_1?-Q?TeRqr1eZ$~M$BP_zkAv%RODEt-`Pt=Rk68)`#WI^9v4F1j@ z|1^A!j)c@%6d`$}e~yZf^!O(hmLyx}zW)3b_>bV&SR2?nv6b;JEK3P2MsOiV(Z2%z z0S!BE1^>i*Mdc`(A7x^aQG8Q*oPSy1hMR4KZ`HOFK2{a{g;%t-l?(hrvczEAy(>bM zjPTzYd|X)BI-@>@ZUQ3L6v2nTLis;c1i; z26rdHw{c8%)DP=6!cQnds2xM-#TRjnb*>&J*WV1qhars~JvLDk6sblxtSf>a{oED# zeO<~}p30!y2-p6s@@jxjt^?;pRqROu8GeB)!GG2k*Zfmsz8z~d0pG@78HdOsVv^i*@kczX zdlhFeukh!}%7NH(C>Jid=n}VJ{sOmp0Jqg2SKQ8P#QN1|9z{Yw-We8GYR zZgK>_yObxhr8u}iZohoLv*VA2MrXx<>(b>+gNb{wh4U7;HnOsyG)fp!D*o49cb)b1 z23Zdh?{DIyiHc*kz;N#&ixT2fC9E`3=zp{nJ}5%4NTI*LBsSSua(#%Rqw}e!8*NkR zf9LIYSpM-HkMs2=OjPW_iE{tAz-Uxcn+4iv3=G%U_!Bt$6DekH6g>qK+n*Sr-G5YxHt-gQ6KI!#m z+upy-yEOl$=U;mL(L~7O`pcZjmcpxxH7Jq!$AAAN++(*{?0|pH+U*+xEx0S06|06ijVLGr4 z@QGc(E|S)*Tf1u2s;REWIN%-55tgD`D*pN8^M9rV9fmHr+NrhoJ~A<{=4Q%|<}BqM zELT4zN))%?u=r!&d*6NR#|`Bt%jf-m`|YPL>xLxwxXu|pX0%(hV4=$=Vc#BdF;lzt zeu^g`>x&%VTfKNkWBBvKjKH(UYKJgxTXy{TNBvHL|I<%PEFkq}x&Fagp=pzY-0D?p zY^U0B3cy?jH%d`e>wh>%8ZuQupT{Uj_Wr^dHvw6S}vQ)+C>ATld(s9yaMexZwLQd^>}*`|IM2}-+q`N z#3d-Y4R|i_`5{b@ddY4zSN4$0ktRNNh_7WJJRD&JQu^DlVVxT^@Jrh*e`p7_ACQgl ze|Y3aLwC8rPXG}9GiS~+{yX;QV-y#IxE`^z1^)Ya=`ZfH&pvalWGPmodNmvWWZj5B zd4LZhQR&1v(F4IBFW~_}xlcHulO9J9LInz@;7`C~gh=h$v~y(@tb13DKV2fG1T@)8 z?v9!^X__bcZw-G#S!gGzBtW^~kHCc=$IQRl`|f8vhRZO7Pl0rhERiDqCl1JrsnZnL zdz|X+YddJI%ofH!(OwGu$El6)M}4nY2j^}#{_7RfYQ)zg+`8Y^S-|pPil1@LIcKX& zS;v3LPd_P$=or_bLpv#H6|2ZP%^@O?PP=OS5mbE`?IPYC@Z&78^WaAW-bLR(+xq_B z+W2Qtga{RQgKy~<f+l z)3|`JGutzwPN4{AZSnDu@K6{ezc6P`8>LLOAt0s8^6j8p=4u3V=9E!>SG`Ih7`68p z*}Cu8@fbH#@j&d8UT{m{KR$ww(3V&sgJfAUTRYULS5t!Lc5Pc)`I*zcQwt=)l~sZQ zNciB&5Q>h|4!Oer!3Xb@tIU5{TVyw=f1qpFpq}L;0Qo)mN5BNf!v}O6|5W;qiCko% zI6xrmDWO{GE`2k?a86Cp`Uh4y7K7eo*eR6KeeBW$$oeR8mk0Q_+;Xd|LV88;z32$* zmH{Y28b}dBN`)WIKe`e4PoL3I8$5*m@SUE~f1MNjN_Ld;MO(}^867BQEh-mY)V1jE zwIpiK_%Es`_*iB5luwE_S>r^>w~#x0;*?+!f}5!O?w1lp9V4iF8R6VT?yk<#Urz8} zdF5qOETD+F_wIWQSzy)7$TT-sSd|PJJlK7yzcy`96lBGp?p=KG#a5PU68NK787#|^ z@p^Bi;y*<5WR9#mNIiPom?UC^KMIl?a6Kj3q|*Ntxwx9G?>z9l_a5J0+QB;H>1Uoc zH(f)&8e*^yY178Y9N-TTE~Jb9b*U7eSUP$7qadR_lqo`w5dw)E;iuxC+xQ3iYoxQ` zzoqd{w3mjz>*=q7XY8-M@=7?s68B7XpQ1aR!u*eAhC<5pku~KmPkC{r=ra-yeb5aytJ>-+!7m z(W^GRi0w_3#4!0H=9$0q6Ho=SW+p0yHbXW)0kR|`mVXk;+MVaztVMIU`@Qf#{^TPQ za4%fAP{CoBhK_iAQoFhZ`0FG{{o?a4OprQ2fo#9~-v8(gIOPTZS+nQb&L)EY-v8kJ z7HoF;^5yckeuEpTkN1*_tAGz+{O=OTTIK_8(BLn9&~5=(vUrKR=bn2^Ko`1d@Ri3d z@tbbG*;TAi-hu1C|9z#KCIP7bSPN)Jy##MAbHIO<0xM%WZ15>!Ra7vl>RU=4w*~)D z7Pw1(USffILrc8)aIrB)F;%`CG)QLk7a9Jq{qMi-w>7J?!T4V;bsiFMO0h z&pr3NtGw4<#TeG$5z36Fl<0D6o}5?kEPdha=)~E#fmM%4~<0pN3ioZUVmM29p+hUSwZ-_9YOyU zT4P>)^%Y?eLCAss=|F&AM9RST-hI!_(vHj>0Y4&Dp5U_%5cqoZsL@t!SHpknjW^6P zV%oIn?)evW2;;sj)(bD${!;9tKZ-`|zXpZCv}x=X&U?E)x7t92R0B5w9qWzX6W!wg)~6`yYQ?95n+lpK^j9^87iU z>#cV{xaPb)@S_HzX7UbyN8!I!_!QVh-#^>>{!jMZQzK6QTM;@(yYE9e)@k0q-J}R1 zrLMmjvT$Ih`mgfeKWdbZNnvZd@Mmg7P;80ij(h)%mh?|ITwzV=a}!04jZG0^m6El% z%FL*yKf%tdri~PP=m|d>A-b*vw@lemE>nsP9|uGa^gqBayJsnfJFJ?r0I8||m)&!Z zZ1CH)Yb9mn8SdP(&(yBsR;mhA;8wRbfloVVJ-~kl@~L z0cs6mzz9;+VBI3T0TW*(>nR*xdD2FS{K6hk#% z-j~fzpwv>jr+LA$vLQQq=08Dp4*XfOXS?U0f8Modey}^?#7?rtm}tTDS~PDVVa<;c zI*)P9bzP-OW%uc)pK8ZZhMO^Cx@&oGEB8hJFWe!m+sJAolg_)h-+ssZ&6m;-+zL`S z5TvzFpWcoi5V)*iXB5KN_U+pz0xPeRfOpteL*0@ee{#)RG;{aedyjT0?P)(o*s)k& zI|29GzmDxfzU}tg+$W!WDvN{zT%X=A*$p!>)v?db+`Zi z`%3|W089V0v#3J33Kkf@@<;%J9<;uEh+96*nb8q+Fd+)gd!mEtv1&(tiFdqKZ zFx%x^QUSoPz4lsn`Q?|n)vMQNH`@>eo$TaVpU~R%>-Vukm`MLH#h*2EhFh_GrGBuL zlEp_6naR(0jT<#qFydKKK9z7STegxQFrVVv7j~K2Av#HP!Z-iJ4?ipy0+Zaxk>8l_ z`I0i_uU@s98$5Kd+jpORj8>~wt(4$)nA^BvqxGBd|M};i?GNjoaiZy(RjXuWbDaD3 zyOCyTgNuOn?b@5dqgu79a^cX(;G=-)(WAR-)AkVC;m7^kZ@um6X{YEuqT68o;9Iiz zN7trZ8xx?TxWTmn<41xY*$G~`a%I=3Q6mM29c%)e%9SeFSWKNZ#kFnQ&hkZ0Ssnf4 zh7bQr{8&~&nOD0*4?Wa)X3(I4is!PI0$a8=R|rGJzjefG^%NW6TLo91IeVtXwt<$z z6li(Xs@3kuBaW1H&Qv#L>J)=rS^+MH4IAbvRIH%guJxrv?C;7~C}(4}V&w|W#}=~E z*x=@f7u#X|BJLja%>WXTeD>~Y6gu+8tk`(ATrl8rSM2W!`@HDPnp zrU%*l!|ehJDAt0^%nWrWpL_rP_vQBC33qePn_aV}P2IpRzjUistaOJRaaa_{S>mQm z8{PNck8-oM&Y<8V&6qLM6jcqhBX#|T^={gfDXv+w=JF-~i{ZcbJ{n7nP365R>v1;X{Bf5utF_TmDl#5T4m!wHC||)9k-&TDFTWU_*)3a4ikJZd`?~`UsB3{`y`>q8 z^JZJmDpaW8W=J70K|I{$%rhMU*uNa~rLE)572{`=loiE_6}7pI;)1}hqehPsj;kdc zY^Ok_rS(H~vwU>V)|{H`RM8e7B8}S zc&IF%iYi$63MtM8i|5x$=}r6~#%{`#sp>~toBz1OV$sJX%822^U46B=U+w*Dt*{KB zZ{8%n9cR{p2OMyK#-g4qdIo7d|Hb%WFRcX=CXBcC;ERBE@?w*fEm)|4n=d{aH*UNu zRJfq5N!DxEat-(0x8J$b&p2H?Jk_m{#m@Ap(_D*|Ewnyos=t%P1K+zkbq>(_P{)0x zwSe98jK!mmKC0dIo!t5VJm2Q#O3j;ZMt*JU%i%{HE+x%0jqPf;aM1#}i)mx;-|xQr z&e~kJUh~Ml4KzMFdl~Z%{Qj_U=b)tV_sqbVhaU#PK<@_l0ZlvS9eyGn zLqr#F^Jf+2 z(%@u=KUT3qNW5RMQlc>Z^YzIISY?MlWXPA6U9Y~Zj|3wtd?%%)4=6ro_8b~gnIwEe znP=9#D#!4zS-qyUM|LXwjEoF(qs8x|&Ye$lhsfdttBf{n6u?~H{cr`9L75b%KUVQ% z^S$EdzY17!51)82q`6pVg$8^F&8-%{5uSMB30WI{xy&82xmP_e61!)X1Ng2E_Bc}(>8m3 z3RodK?s4jl10u6cjsMqyI19F_wIY{ zb>k$s`?%jna)a}^%goF)0bdOTz4mzIGhy~r3CiC8@B??!Nhe8|h;a+sYegF(pm2+Nt?Yf<+76efQlfrOkAA*IoCx6%x?B zqn+-3diT+o{b*Jv2#0&>hxE*uvl8%iVRVK51a(K4|Kt--YV3xZQU!V(ckHq5fd?Km zA?^eBKVZUk6eI}Gzxn1HchB8-xs97Qx;Ni^)5h+TPd{@*r5JcgJH6k1=UuaMx#pT{ zE#PpEn{INySkTgZa%r}gDJt3=($*D~QVCaOOC+R4X*5AfnXfhW1f6_RKkgbfXyo?Y zw{}7ig3$dw(FXzUbI(5GCQhE{ZkKWmp(9F@C*?C* z^s>zXb{ZCt!1s~I9yNS=Xdbj^(ZW3=Yne|z?&sco>rMBA`2LD3u5=%K@S%jaQ_QW` z|cMiAPE#cwxT zkZ|y4*Db-Whgu7TP|xwKx25Cpj>zaDw)F)7Cu8vn0dyDkg=A0#CLJ01xb zdhdPriw`$he{Pr3#q+B(1uFBQQDYQZ`6W=EFX2p2w_FOJ3opFDRg?1Uaq-Vj5>A7E zIVl5f7EPBeTkfvC?pkxH^|Jf7|ZQ%cSgEvu2GuL&~`^QZ)F) zJQwa#CcPmm2H^j;_BZQ07DHEFak=~Pryos7S)qJIcY^|fu9Sjejg&PXy#Ik&j+K!D z4EW{Bmv^_xw>CoffD>od-*w`b1q?sFXHWif@+4V&EOk1C*UlKl6;MDn$kJ zY5e$!7GxF0MX?gat^QL_KVjBmeZ-$Ab$#;q6A*sMk4s#?k3Mo6wKlSgl=bJkZ%4W+ z)vB06q(}Fg+y*I9u944je3knplKMUoymhhKK(W`Q%h^&QA7tywVMiQp^S`=;H7IF- zzfellPe1*{Jt8IPE3dpN#q}pfBU&mTrF07^J>cK{_p5EeYZ3OmBL(`4FZ6aF_UmVS zfWn4#ob-|U&A6d(@6XkoS*WKTfAW9T+VPC$ zHVQVBr}XJdDKn=|pDLd2De9xJQw@!kq*A$;_n_~xaTUve{M&eXc~fGKpIc;YG7P6{dJ;f{L$+uKp( zihqAZM1XusRw6nluXiT=UHSd9qwvQKkrz5cOe37l|K75M>Ka*uZ0Y+CMJDUT=uyO^ z@cc{P{Pz!6;m;o0co`U!1qRYeE0g@FzS{E&T-xEW)lf^S6A8I+Gvx)*tZ5Xo26SL% zP>*B$&mQ{rbcwKRN5KbwYce|ee#T+@e1IE1R(Qw3-x~g-qj({0p% zK$Qr8Mlg99m7}B|UtixBU=-~tDvu&$kX&OmkRl|sN;Th}3MIzhH=B5Ft)O*N(!(Tt zD)S#WROW`W+KTuuz|YEm){w82vi|yZe2260Uj!dWK3gyhcoz6`q*Nr;*X9)XpQ7;l z*5J>Xo3uVVUY z+i@p1RKJ5RyzpOAqSujPy`=)Lzhbq#)>D@e{pZL%mL-v(g*I;0J)n;Is6W-S{;KbH z-f^elPYje){E0F0w0?ipuh+obblq|1UFLq6cJ5F;ENpV5e+1uPZv+0R@ZlBXnr$8b zH2jEvGy*%B{$WjqYyBUi9-n{Q`vhP66wQCT2@{oNA(d>CXk<_dh-V!nkwZ{`XH_*8lMR@A)`QeW6oKIENpmy!o@#*G=MH7iLx-;O!## z1q8Q-6dAQ8s5LbxP70K2?#ROrlMwwlBSaSb^`T5! zv0}M|(D}8~UG5lkU$sgVxxBhb3XvXW!N91^m_EaVmY;t9xd~qT_a9(&bEE%H67sQg z9_t9SU<3PIDA1RVCa# zTFQVcuDC*i-vuVLuX|uUhowW+s#WdBMxFiZy2Vn|U{z78R))Lnwp-m{9ooz4qntZl zu|M{&Q`dHc8w*7EM_-Bd~iF2HDqkHt$iYwBoVURzTP>_K`bd<}BM~kJU|G{Xjdo zMJw0&w9c-Te!yX6v`oI83u$a|6~fN-di5K+3@O$QX?=(*R-~AAy>G|@KhVGNL5)oj zfwc*JC@TRrLTGkq>*uEbO-fC5y`XscCwepWu?@>y5*d6&ylr25#xyl5+>x=zwxpT!@>br^dEvU2L6D>#wzjE*O4Vh zbGJxBXJ`f82M!t}LG-?^h6MNsvQhB-_S<^nzlR_Cj|HzL_Rc#}A}#%8sl~yVEagD$ z+WShm_khu#xe5PanT0h50Yiz~g9SyCCQaPgXP+(H_L49jK{zF?ccxgXtaakxmaW_& zQVM2dWLQwvJxZ2#dnkrVlg0;0VK-Dt-SzrSxyj9xWep08<}F&Pk88vy%iP|wz`Fk0 zYs6obtnF|B0?MMc5_Tgr!E&s==!{|w_CZmC)!c+h6J=dC)>V|{32`p)dw%+9XSj3D zI@_9JNcY)yAGe?Wu&zOA*|uE=Suh-EZM2o*Z=7NsVHJa*9|bJ)gg+LkufP7r))n~8 z2Wd^>5B&=iDrobcwH!Br6)RTMSbgW7eEdnH(o|W-)Ty(-d+34tWzA5_#x&56*iT<7 zeoUrz!`9bM+2s-dG1lzl4Hh`?B=ZjoLgs(huH9T`DU4BypM26OifeOtVjc$mojZ4d zSx{iT)Ly%+r%j(G%dSj^HBIL;I!h^3!f3P4-ut*$`@U-UV;RT#2@jy`+awDo5MVx{ zWCI8IWiMG1h5mzoA-OZlFXiRUH{ERb<2tZrjZD|^xQ@1l8nx7p@fa0!9qQAGg87My zvX8Dd+k1-juASDtdb094SnmVo2xFa@ndykJg95yw)*{{=D3Gu`!LkvhGi8nv@3wD$ zsL3#pV_1R*g#V(2i_Fjc`VE^L>mQbLO|%wo5`PXKKFr1dVP++bp{nGN{L|2X&>@9$ zp#RpsKSF(9*^aalF6_KB;b(dOC%`5yb~^lR#h)O}y`FES7ZEW!`KN(@!n_OT-7)`& zZNf&KjG7sdYpj7AWV4LPLMA$qU2)^+$h z#GQSXtTvQq6UVnmEzt#fh*}X{;`k@lddla5Ki`yCMfdyo6IpFstZ&YmZmhl?{-sSL zY#i}40D}LE!>50||FI%$txX2>y|wz@O3%;KX13#T@0A1o#7-HecgqOie%r`Nx%>tA~Qg}LX~B&cI6{Wg66 z{XO3QylvnXqfX+z^XGj3M5-T;3W#ZmwHufq`P8$7KPE7?(*6)i23`qgpR0ON0c$~r zsHH>;p*|nGy8=GD=5bdAF1RxSer9H-`%W%CHfWb_)oN9}kjeV0g9K~1AHmNv!Mwv@ zv)g>7c6|RVmr(3PmK}rnmfs{{c1Z;+$7%zm#_YMX4Oj3+fZgY%KKcuHJ5&+|l2LIS+o^hsGA$+eL<#ZaiUANwPi@Wjq8{Dl@ zFueTIi>j*_1rb3JC~QQqu6tZS^^oG@f`48hH(z_pvS6ZHqg}|eW&U1T?ov)X@kDd` zk|7}#_GQ})_mK!CBN6?6Xu|S~$iW~BixdjSLvqAkYE5RxXq{4*? z*_a_@U##8a_uPArxsxd-7c)OgIJ-gaRB$7-L4Cn3R`C+Wq>!rWrc9e+g6Uck^5KqV zBV)*jXdn%4e*H9v26dzQ3*7>Ic%_JTup_KCmpt%*{%E6;1kyifXEcJ!ij^w5y`|{c zyB0gLe^3DIR&w{$79gS@$_?jUmIVZ^oGQvq)z7jFATV?se!c?YpcMGvzf|-_FpFy- zh8Fxse*2A^Cc%EAMh)ExSrx32Fnh)FrN&1rPpdSx_uPB8JNe|3T&dD~mldgpnKQxV7GTbGZfLG8u9)G+ktq1^G zMdM|K)GL0}W8qM=cnKRv+)EL=VxN69j#5@^0vR=!=)dTU+oJuX;Joaz%M^t328&Zr zUv9;Q4jrmk4+XW0@Bo)79)I(#w+t7QWy}M{A-@!CVgAGaOXbrZ{`XWD9Q6mmG?p{C zZu?%A87OyZODK*DBjz0fa(3eXsxf0u;L>g4qzSf5|2vgONz<`YC;7A<>JF5mtf<^P z(I#_xzN{_?pw5^Rp9Ccv@JmP;2>+Mfqm=a(-o!#>FDVkhYrTZiB;aE;hO4f8@HE8c zsE*?z>-~>DlmMZvt$hT^#7!6qeNTE2^L$}I$ z3%4g&ntU~ExNFk1iLBUG8*MhO-yru^d&>RXG`V=2;%3a6Y3mOGZs9fH!pA2_(A=d< z7x(r%Z=1Wp{sa4K{qXa6ljbyQ7h|wd>j-Y-*qKW{iT4k?+TiET}oQUvXa(9)*;?0^oe&W zyanwCeL2|h27eTImMD{w0{sHr!`T;8`KW-^)FHKnP@nR<;`<|4`0Ds>g})8>N+aYu_%S zQADXgf9Mq@qn)k;-NV^uGdPJWBY2>BR6S(dZNNWKNl$(Gyu&9p9^boc+~NHc09Xq; zL@fpIAv;!1B}|a6j_ZH&{ZD0T`{%x5v@YsVep?`I3G;318PMQAlCAbMVNVm_V(|U^ zH~!oP_+=uia=!maT0%XJe!glI+{j6D05S08$SySMC z)#BX5@DF2UU{dfmjPr;;ik5rtz1MydJteC~>h|sXs=b@z^e+?j70TVC^7QR#_4S#j zbk_B=&*)lxeOA|4fkZE?9PzIw>)uOM2G^^&6h#q3PwkLwYwCz^;_qLZEk>A~|Nc$WKTZF?BmaeW)nDcPSD=@Yy@l<- z&reXZH&XOIjy+QjyMO}nyq9_LvG9Zi!J5f(ptP0n^>iERhhhloaFNnTKU_{Y;RHPn zjZ$LA!bf;LK^6@NxeG`*I!C70nVFfcu`EBZ2*9mKhYlSis65d8>lZ0p$byfrUbV&* zltAkj{lH?EEdJJuNbojq-aHASHn<<8I6!QKOAiD;2xEuKultV@DxP-QDJBd9Zv?gD z$4!tR^?UQ5jzEamB>X7BUDF)-y2V8o?w-z+^}<3~arD=2dxF|xebBT;b91eO5Vl{x zPi&WVc?qND&z+|o>OTnfMia8q2L#eTE12{o31e}4!>*`K62w-mT1^744HA~UW3e8< z9~VNnGr_GA?wQzG{n=-qDwfBg7Hs~9MT@oL{y=x2c97u;r>u5-pL0%E(d%P#Su{`m zzWUm$%}vYv1q&?3zyh6d|MlaKKbVr}yz{!rwbI8D02fmnl)4rx1lKkz%;lnR`?_Em*@){OTprZQAPJCRkAY4JPMx2g) z_uW^*_#-Ty3gd9GcAt%x@E;}5{&n_q(`U@E9sSUX-L89R=RNeVB0(og4TQ5_eDS#n z?GLP5&&KFyjsIA=-YJ8-B$?Z@L;rt@1yHR@HP=ZoXg(JI43mU|=YLUm)oSXfQ8)s9A7qp_p1uGXWfvW{7| zVVwndM_`ZA1#6Y!QY4|U#RUax-(u0ZxGWLMlq+KjGpwtKsZm1f(6L7!<$Cw-EuOC8 z>S+#MeDOt!{c?oXg;JvHPp14~!N(02>;6FpHI;j-cU(ioqd`#LL{?Tbqy<;tO|DSll2c#afE>79Z*7oO8B#W1hVuu-HN|!q*~NUa|Rq zxU5u&O*BID0(W6YX?*8ujuShoT=}vVd#0k^9axW%csHReAyx;mh|0<(Ar_M;_GizW z?eb}SR)_{DQBZiZwxGl(#t-Y!GL0!?hmx&P!$#JQexisM>g&C@*5b!>Xy4xEJIcIw z-hIauo2>ot02U#MZVgPwc%c2pJCm}K3kqJQYWJM|vn)>`La8v#S>)CE@l zM7vQ{Jyk)9MMXo6T@Alzu_9WdO53<|-qrB8g#SesUu|w=dujs>@qYrbb_V~e@hIhu z|H63fsQDM5FVi={e>4xVM#Ft9^kGa$%zAL=2O#j_CB9=k|0y@Yf6g`)p``bg3vE@J zOqGf1S_{^Q9>Fte8yy-ntZy4y62)|pI{tbPJ@feIHvVC~jMG2aZW8{s^55SH z{)hTqHcoRE`2*kJVf?v(_ZR`vUwQq@f@?JYvpn00`1Ae4rWll%HXrk_{%q^~OY?u= zVp~|kjqx>y^!gK+Ui3-xpYg|*-v8q6gC51NlO@8H&)ssFP&TxcENxfOEdsc1-MYE1`a%6d z?-!%bj!-ArM%F7|IoV>?+|;|&(srLiWdzPB9!NQ4$C~xI>}CxvRAQNSi9jr$W($A5 zbaYf5>JbbaWZ4PfKe)2O%oRRErg|K5EtArj2(r~7i+gEAOu)q@$@(i{9Zka)D0i%n zmYX~v3`3C54q@V}bka`b(;zZlCn{_x*sDb8>!6i=z} z>7M6f^)~7&JKEV94lOB&Z#dYABlRY{2!iM>*_mJB85_UTF*TpV)PG$ zp#ddmz#q5=LAU2?TJ{*(eKh~6Ls^tUaeSXuTE>~kKwafm!_uiL6 z>o{4Fl#0sbeDZB+ob=DGBl>3{u&UCp{2E(=zq5P)jLw}cZb|l5X$8vx@s+N;kjV2O zxPuGHcn2VA(?j<7TW#Njl?w;jGjb}=+T?6B5T6^*>0W)LN!FkK0|v;gRZY1&+%Gg? zFvAUe$UnvsWsCJ$_2IKkQs8_jh1mXbIYl6H5?wNDHeJMd(K53f{P z7$^k~YJ89>vaDa$O4P^bI(~xwPdxdA6rpR~<#LZ$i1(%FO#e~tqTo(VHm=|U&np3k z`Dcrgx&+3gIDD++#9+ot*=gWQ#>-m+xg1hU+kFm|v$P|o!Vmd67eUgrMVwG zaFA=)p}pI)Oli0D#~Gd{^gMO`2i^+LX;&+(=oDK~#6Gr6P@e*cl!Q#JP>1X={}Ss= zR42d*&;M@l*&&~~cVnXl-!MT~9FZqFCFz=#-dRh9yVTNl?;nODc8))aUuP*4E4k$| z$y&Z#vb0?#Z>?aMhczXH#|3N{fB!9`B4QZh8C}AEa2m(({<9n1U#JZieYIhvd-v{+ zm}~Hyt>r2Z_|NL;01Q2(K}(zh|Mit6@rD1qz}#X!_0-c_gn!gAN4vqYq(h1PKe-%@ zCtxm>)$N|zT#yR?9Un!JwQGWvd#4|Gsk`(-j)4^u4NwTb{A({mo zfX#EPj0DdOp=gZviL+n1&#U3uS0-;;RSxoT5Gcealj}L@Q=gK%lHE`o&x;quDwP| zt0xRL{hlJaKyP)>cPKig(tk(6kJCRDf9urW7=9T47bT16dhWS4e9(xY$vOGlEtg5w zOf7BqIaH26ieD#N3RgaN%Vk1Y&t9>&r46_7atea+_TtycmcrGJg`b(3=~#d9Sb-9B zXUxAe|0ma9MlZhp{oUt(8h<~pZC0ne|I_autsF)e%^%K6Tl)Uw1T07B>(GKErW=WDjD8}^Rna4|*dRdrS7R1E(C+Qy-Ec!ax;rr&*7C+7+ zT3A_MK_Ih^KZ5&2R+N_;{A9VjjXy2tiJdfoNcEBNtuawAq;JJ&lUgyFt_LO%Ke zHTZ!zu^wlGuLmBR7*2MHaF_;?HF#w4o-hhZ1$bO8LZ;sG!~H#~ycV7XGvP98}IjL75XVk9C$7 zljCni6Ni9J;(_X_d_E6f7apG!{}9 z5DE4Zbpj2;13++X;6!~1zp?%}LY5jEWhFrdaZOG>`9!m{qG&Q@$KSUF{Q}KYhK@!l z)Cm0q|IVkMF2S}}5wZpR)t>?}n%JBb{=SAkT64PeO1;glMDb$U%?J+^Nx-9tFzbWIcvhVW$TYUj32VYr?kf|20upYP{Wt=ZTai$_=oYQPdf&F1XcanJ@Av% z-Ieg$wr^*Gc48e2l&h?EZCiQ#H5fJ?cIH5@G!7=6)F|4Hd=$r0YBGj;jv~(Youd`- zx@FR6LYZvroQZi^0L&Sf}WlezvUnn_t^cdA~ zvPS<^Z4J_ua`eA}ERs=B#o^y`*FB~X|4`PrSU_H)7<)BqWXN)3sadK#@%WQw>1pn+ zpu6D1|5(sssW5o(mx>A1(mnOmlTymow4P#x(YNo*y3X%D{iI(u{K3bXqEkjIdaG-j zf8m;Heva{k(Y))fyWH)!-yzs%YqQXqvK(w^&&Q1&Cu@?&++0~@l8!%4F%^{)r$0N# zKX~8o;y?J{7IKsIv@8ZQteo%fE4tofR?pDYAPWA6s?%AUlxiv-7)mNEbZgekG?%;t zY)5%T--(aHxP#92kH24k;_z8x=E{2C+Euq0Yuw+OG9(&*xD2k^S@4CDTKD`H#@{ea z!?$r^{Nb38{IyZc#>Mlta>F1(%PciQO)C&??Cjg(=<^t7|6=mRLXE#O{6POS|3_Q` zQ;1LE>z^%YTYdlJy#Dys((~VcGiqK(-+!#rVV0-Y-}L(@z5e83{jn`5l0(E_SpHRJ zYgXc289nxNaF$2#)?ootVS z{E;Z9ljVnt@VQ|RtyK8_*m3X!4Bv$`e8{|Y`WKfgpR>Es+WfueKZ>Mg&70X@pcR3# zizX~re$V`6eU1ly95IV+Way3 z52Z-aqF%|D91rz1fd_-8n_T9fr$J&*M66zDAUpoJ_aR0|DB;;++vy+b!Dk8TCK?SF zAumxPu%+OKJdb1YKJWMkgcGd?unFt}j6{i@03V;0G9j9l$0oLf+27 z|H><`aJ_on;vRYIY1h6jF{Si#RG`@Cz=emN=oB)%h?S#foYYtQIZHk2Q%j`u(o8+H zO1kIEU4jIkRMB8W-!0)6P#V7L{nCF$8)q(6x$qPVbc1xgWmp?u{Oyao6?ZG{TA)a9 zTD-UxFH&5J6Wq19yL+L<-HKa_yHnf~2+7IsJm-JzeR1#mJbPwKX7--%TA$Sr63USD z()Sx8qUsq9!nY3VIfAu#(*)}htN8>IJFsEqhocbsT`^*4r7@Z}M$?B{zSOf7A$|Ka zj1-&x1(8avtcE`Ne(P)SoGd;K61dOh@v%oGj6j*qh2iS*@$uWYFSg!SU%q{z)?4(K zU@hDo^F;#f4@D-ihq&h9sh~hpOjy=4Y_m~L@rt`|Sql5%kAc&T4nY?Ok|TzY!)=3C zK@s_V6Ta&)D*>0yKN6Mrjg~cHap7 zj(BVFO)0+P7rsQacy|;)U7qbAvXQOhdim??M?~P#03|g96L`L|ozJ6@ytiu+yqh-v zi#&_c>VUDo?S5G?I}*Ipbvhd)w2AoddiWjyobD*_QTNJs`(d!l0#PNOUyhyC;XUdeK7&ticszmKN zn>IL#t z$VkGjLSfpKXoYDX6QDRdGoA6qg&w7VGs)$fG5%?qil89GNF~FVu=1gw1uB)o?S=1O zPmj)tFRAoR*q?wBCNlF8RNS9hXX5yv#JD>$OE%KZookly6uC3-T83BBH8HnNkSI{Di3KBqz3iZar9gl#6Kk-ecWEby zQV{P_`Ioe&@kY$HcYQqI!gUB6r2F}ky7UywEg?l%_aU{x-2NAN!Thgjz6eXG54+AW zF(%O>^xd`9HAEg=#kK^N((JYGX9>-F{2u-+$bbnnKHO;jz zVC%w_#=2;cKG;s`mJ1m&#U=SE134Lh_* z44gMuJ9K0AB3#zEZwLGX<6pR0XL5fj70Od5cKZ^OaqtooNi^^Xf2z_-k9Cb|Q9k(< zyF?wXWT|Pfh}Hbb?&m<3R^-a~>FlEb?LN*9XckAR&#$NNB|5vrhX zv#_F07wFA&A9XmUBTabF7n>%sl=j3m^`j5eX2B*Pk4VG6ADI>pP!5qm>?^>;Jn4 z1b~{lUoN{Ank?ShZ-5lb`)sPQWq{9HbV>AAu;~Kk%^j)OOFOMI-~x;dX~kB~)vKa) z5D;m5!s&~Hr1l7F?bEwDA#&2BKUJL_=(K*OesqVc^i z-+l_~{Z~^Y9}_J&s4stG^|JlCmQM*)+N}gorP&NA;bdZAxc_HI)R7L&2f*mU2KTtz zd3Ti7H!-4}KUh@TXQ>0HtDCwzk3XVHAuB=?@n_L+seUh;0l8--U+P$YMbxJYHGvA{ z5f0xBDcNZrvA#)L)Y$Y{{f>6RK1^`3+(x%&KQ<|MK`Yy1Hr_HT!HH9Hsrp78-(mW< z$_xzW`*?39;EL+aT!(J$5?LU0pGPqfTt1ocu`lSAnAZ5W^^JxK%=r0gvrm63U|-FC z8_IC*SC>`?OQH$8#ayD<4!BDY>H)LhIykYNLe&5P%s=!ZNOxgw7=!G?SI=7dxS{I$ zG=6XeGSR>Jw@L?p_04pP-=LP*=L0KM!75*1A}UxU>{YKCBA>?+L-C94RKCK9pqSSP z=`*UQ#ef?C70ehy5#DXw=6p+ey@^c+f&0*g@cy6^|DuA+A*C!X`#J0v#vZoK zU41HlTb^%pNb&3ex!U?^?DufVjL-0>H=QFKO#!8XvI3gAklsLY#$7`SDyV>ZgDcFH zMZ<={!Ud)&!d1x9$quOUQG%4&Egg!c1{Z-m8*&b8hxb?{P-3hN`@W;2ZErjK>IRw zJsdk0fBt|{Liev3U^ZotF%_Wc6Yn7C?gqFAUcT*efxH}^!FSs}{TE{eDQYW}-+HnR0+gCeqWpE;}dV3d77}t}#}-yc>_UzpdA6 z2?)RyUXehAaKDsKDL=kkQ2MsZ1)-?%{0|GjKU>tE@tfbk-vPsu!naCz>{?~f&m=5G z*a?8uTKRFpA3>W&^`tSwp{#i|M)Rb{t2T*GW@I@tFHAlj5u-LQ6oDhS)9A3c& zAJZBBa;zri$}-!-qZQ?a$PLNbJQZwCOw!PuVgRQtY|Hyb=EVaxvevrrjM~fSz5e{( zPEH9n;I1!2LNfSg>Vn#?E)FFP4V3t5u18`mU|u2KVO)M348$2c2r0zZ!LCRf1xj#< z^inXS8*B!HusUj5zak;47BASE$F#( zz@FN?E?^WJT!CJMvgjEJj*9*CTng~voB0b7T!dk`(%fH&V643|nat}hXAv6za!xM{ z0#hupReG|OMZoY&0(C^`WUt*=CV|JXu#-=vHV-+%mspy{?odv>IyQ6b3?j|PPa(d> zxl35fPy@+qZ6H9;4MFx>7l&2`diC7su>Cg zVv{xDiJ6C&(xYr+fXC#2+|2Ia4>YkNiy08DL{-p9ifE$E-XV57orT!AiW}BHb4_Mt z8%7{w@fv=tR|<6X?cdq0MS;S-q1GIhZ3Mtkh95DlEgxqaM{m$kE<^_!unEG|ne8ZO z&$XtCaQB-{J8)r4o3qniFm41(eK5(CiYV!}Eu>#6@R0lniT`|~bd~)bB1r|=^ndyd zpKp4cgw&nn0C$y1J-%Gpx9E>osi`a6L9oE5$+I56kG>dg$5Z*72^PZ5J~xc)#=q)x zhIt?Z+X1!5wR(*il)^QW=Y}3-fl*OqI@JpV80sk}Cw4L6NWm5BU6$U;TRQ|Dv?0)? zIx}9XWtZf2xW^@f{=;6eTvWMl+!uj|EBhY~f0ccApfB+h;v1Y7>n-|HkaoI%7xc9m zjcF_S;x;!6%nr<26kU&*L&QNBYhIm;z0x6}(qb~E6?3?=EEwN4EHmG|A{R>0@Rp9o zr-$)nnsr(9cm|cFT1O!rw3z|ZO;;35>iu-RmSxNP`+PpRtW-6dcyD?M;OUVPcQ1DL z+V8U{@ApBURf4F5wQf-kBgHd)PUmhcPGRGf1ePQ^+ie~0sN+HXhWHZV-hT3UvU8xP zLf{0nJ?vycRlt6N$*}E9{RXmGAKl{xr)gI&(;_sR1AQM)2C)@geAQ^8zp?kb$;u~q zebd_1EecQhI5Gc^7moioq%;ZfUrwoBXUiYEqG-Rn^gdG#{*n7mM#|?OJX`ttQk$Gh ziZPkVn^~?xS}O**(B#4m-jM=+i2hS$sO{W$@oog>0a9+z`}rQl_op?P-3G89VYN4~ zOX9|W3~Y1I`0tiO|1nRi$4D5wGt{fa_@j(^rAuSW=MMLrV+`->_4W*n_+hR~U*vj) z51S6WV*pDQ_EoI&6o5aB``kIaPE>x^fyLOtE7s!)8GZECa0t%v%(5Ba?WyKs$u zDvQV_SKsSn?l6P-7m4sL(*hrpAUa9C$5P(Gw}Unws__uSzJ)ity8;gPcIOB#l|1zo zvmmAQo9nkpA4D(RmS1)l!feD*M!LRT&71E1b@-F~wB@TGOU%t7@)pBebxLo}hwhtK zVFoMh6{9x&RHDKv&~_$`C4usz*^Y6uJF}SINwi}mRhZITWnT%*`zk&{rMXnA#?v1B zux!SM=I~)^iMy#utbPfRroSb7PJdFrAvv8hGYl_j{5^*1%OGXR%jBkXmq@w9x8B)Y z5i`HOE$n?B^`q(hu$}L@2bca=sZ%?^8PdKaI`qRWqYhZX%Rw0amOgCl>-<#<{htIx z8c}qZ$Qb~tpGX^MH{fiv@_xRQ^tT2tls63(Vxv zB1^rPge}aM&jMl78E~-ZN76YMB2Ev_1&l)J9320nelp6@NFYIaAE^|=Xw4J zw{iOJMvxu2+%w;lC3jhtWBqU)kX`*{g^rZ7&0r%f>5ii9ALWzMcj?i)$bEmT6eksT2nY$e_N5D>=q~Bc-Z>_#{f<*K+k#_0 zf?IS*xj#qKJgt_~qs*JmX~Yj$e9>ZLxt;O9nz(s8b|6Ze(9yIGK#=M5cHFFh0NnPN z;)@Mde|dFDtTfaIJd9*6wA*X9AKa#n#F17rqxj}{V6A%~d5khlp(aT^W!wX*J)lBD2p-!VUoJPoD&BOKC@^@J*07swy$DHRD?d( z8@D0Oj!d3xD~_iDiV3)Bc>w)?(d&1W<>mX|^AyhXL53X`%9yr zv_f`G2iP;Lz*n#c3Xw%kQ~jsfUo&QK9DX?ADQ))ca^a7!N_F9~*=4;GKuU%Yxgow| z>a@y&AS0<^@>2|z$fh~X1XIi9LX^6y zzx}q-ysyOX5}zj5bZ;CUWQb-`*%wR3 zua6p3KJ!-R_8id^a0)>1qS3-mDK|LI7&PLi4E(m9@BLdkzAIp0AlMMgEp&~2W1QSY zC!gNaZSf|Desfyi-BgmYu->;qqT?trQ`KR{7V<5we-1 z342EIM2u%$UyRjNA{(Hz2}ygDK*zAfCQ{#uLfsrrunx{&Dq}wkOYY z4ZHU5x#|7>3%L$==zG}O+M5e~@A7;E@Vxjd3-y_bZC;Ce{y63II<%ZD+ z7Cjm}5XOEp6vT~C)H?%H)g0^X0*B=Qgu2$arRG{gH+)^igw@ieTQw=-@@G9+<;mlpnn;lD zzqzWlqR1Y>A2hZ79iR3Uwc0Kp2#j!3e-$x!^WaBy5qAp2#&d9U=ZNTUVR934)G}P- zaO>e9oe9e>^S=$E4tZW1^ebWvzaL>@tZNqvZu%~((Rt0G--%Oznu-AZ#>Uwj`QzX- z#*~NbjQ~lgmXv;ro*KJJJ4PDO>OtcARdAE1B(YZRK`n3JU}`p^ZLMQ4fsc6jFOeZr zO*stH-%;SZs1Nq7jk^iBOjO^hZNnD!AB|mzZk48f)yr~5XfeM%v^{@6Yt|Hum9pNS zPZ$xb7 zGmrjPbYJ1Y?+oOw_M{Jjb9U4lG2Efjfi&S@{GlJ=ST(38Nq|vUvFyp9rln4urp!bd zJ9a^BF%~9P8W41Y-mx2rw$SaX!e$!JNsXq9Cv6g&s_8k0bPA_TrKFCjmzZ)Ed3KrY zw&1`2c7jnjy?9`D{b?H?5W3L71CiI)(tqDdxz_I7a}ZUOC%aOwxu=>fM6q8%V{)IC zq0FXF9B4>^HF&en$_CS|lP>(tG9nutKz!*!j?AExm`B(nMTxXuxANL!T2jJTZa2>; z;4!KSb_3yS32-Rq2=$B)!SYDm52u*F>NOz6m;BSSBjbwP@;!6fY&N37;{KsPKy1qy ztgUD8VR+avRZ-;Kf&ob!DR)T8IaT{6rp~fsE8{pGbVp8Sm{=b8&dxPV{Q$?(7aO9J zGzQ8S3M!e(mNZVx7v&+MmjIweplZKq4T+s6z;JJqrD#H*VCO);ldV#V<3i&)Y*@=* zS%s1vVXDaw7ArDwbu+$YBso&W@K2%^JozL zaG2FdR%6`cMkbS8qS#>C1ZvBc=W#=t!}F{l12)z@*BLA)Sn~sG63ueF_+_^?G?buQ zchZi8v3`BYw1pu z7O5KFB{NxCRi&j>?yeg%(q0#S#;6cmqs56m+CLma#Ehq$wIeYQt+dOV%BrSy$XcKN zZ>gHIx$B$kH5tPpr;e7*CssD%G1*;FAT$7+u#XlyQ3wPpxF1ZQt~qOEI&kXOFEr17 zhm)qmFolEfCP}C^AKbNdiXkIfX%2Z zKlr+BeQn<)pB}AMe!|XH&d4}ed@mXe4w9O1fTt;C7oDM#*y#ABlD%_U!RW1fE6g`W zxbbFcT!(Qp$+v)%>zrt&kA&}XsEBC-i+Q5nX|1Z{>+xeg974*ViE)vvmMb#(V`S1A zJx$K5*wF=XB$h>Itq;bkiL{L(>g6>gdBck^f^Hh5AZMXn`UQ82UzZP`M3dQYAT+|W z=dV{xI$PHcy4Q^kZ}u#jf8)$EwAd#9J&0*>#0SQAu{bK`Bow9c08OQY#EU3h&U!%dag-THWmsrTgEl7A^JNn`4?_|1=_TOJ6{r}x)-S%nG3O&rxakn2_ zkK=zh|8nu|x)3UOUPdz`)#miq!-)@sETQNucp-wH49iN>(AovSd7O0_s1j0FVJ^|h zryM1~jb72$ODokhHl9RfDFZF} z{>_DV7~-dQ-R8xzT2-zVB(kKz*J&X|9pa+s8_C3x=CR^Q>`hpa$+x26mBwu&{q?$- zr+;fRyI&uTIUH^0q;ljMLkatN`|(4ek(3*}aD;bOQFPdKP$|cU6TRs66jB_`u>7pf!W!>9xv1nO$zT)r7>SR7NLerK9;|&SBdyW8N7{<#fp0yFn z++Zjc0b0wFJHkrYM4K@A&ya`H=h31gdnUA2zc!Qdg(YC|P|8mY2s1U2GDd$=HxGYqt*y%5k5{ zqvNWu9_l)^5tnLgXknto)6=hu^@Ei+!)hUW?2L@EFR`(Nql8&#%p^o-HvKLO>7>&B z=7(Vc^j;&bAanAU{rg z@!Fej@$aMPMK#ArBmQoj2117U+357E5%MJZi*m7acEc=!$=0Q@F|fS7%`3$teuL$1 zzxL=z*vMwr*%8?r))~AhAHa-PGIDywe3Cp@;ec-`fU(jkdMNu4HezDqVwAYhsf zr?W=v&Dl3WM_hK;V3ejGJ9)8`9NALmFU5RN{FemfpBhasTLC9~3F52PJBA8cJp)X{jy1k^65z~KAS_cUEYdz60z7LsMrSWxWMO6O6xR^{DfHY9UZuY zj~L?|XtX)^J|y;Ma}(n}XvaXH(W?XtJ2B8m$T5;W&$$I@2dkPX{H~R{CFF;Fwo>Oi zmnGFx4HfQN3L@^wlhARx{S+kd-x?GI@&)>ja+qRDGv{kh2QF9>_D=I19;_Hw=%8G&6v+8 z9h|(;8J`)!)w#k7U^+&4jlK9$XX*_rU*>Zu-o8NM<@QpeS>>w+nPvDa?xT)6&*yFw zr|YmEOS>e02GtNu#$r===Z{~^Lu9@deJ95s&W~{(SsI89JHtFq7lwl(rA{>9e zEtp@iC?Y2S-al;8vd5+JrpO`ykdCg0oVGh1i-qdH+lbXU<+!QJgY zp1r)U6voMSB|oyKr5Tp1Un6%)a{>`}&47WXUU7@{N+O$T>)pPB=UHa#eK3A?)G1H3 z7MnX#5~g`G@H+!id6!@tM3iz*jgkD_+!B$0&d2bteX`~>-0`NL@b#!#6MZ_2MMwxP zgZc86lT!(V-u2G}b~^vNQN};ls^F}~ZIX(Xkv{bOL99oLzC#%6+{AXh7HfLG5u!SY z`56eAS8_%CiGFD)d&p^^_@@Wz=lidkwM;2V-~5wh`0Wir5x0{0(e%ouK90C&q3;p! zmJlh!Wu;xl)DM?df?Rgl$sLZTs<9bHzXHHY?It?BGK`(qI-HzJ$P16nyc-pKwI-^n?Ir*{W8 z0ok8DTd=yuwd;}~WaP=Zqwy4_Ja)gr@XqK1%KxIH-fPo;sPw`q%*BWcp(&!<^tcFH z&*gR{@_6un@ds^8I(5VI;~I7GgvSpIFi>tY zf0@)H;n6bhCV5y~vqUYhrGGSuZk&gEQ?dar1>N_1Dk~3tR7l&@I|C#F9!GsU33lxQ z>zukZdMU@n2xp6xNu7B}4mmy$vP1Hf9^!8^+A@S)oO=5?$#m@0P^%4}L-31vkNjB_ z9H(czk2QxB5^6+54E0_`=v+w9c66-sJlmpamX53z#@n>w*Bj;qydpg*&dPlxC<_gG zSLEZJ04$yOs(OcAdu6YJFc) zG3MLyzF0%;bd>9Q7Z&vkp^}_LYb;p3W9}(#C+J_7#dfs~;^sjLl$b!{^()8iyYVf_ z7xSM#RbQS9PZw>L&JD?vjCBW8^pyaw5Td$0(vz8>TXmt6&M(srhaZjPB$K2})DYW- zX&zH^33k%+EaW zl`pA;MJAD61a-<477c48T~t_U59CTDe^K?x+JXB4a>bHkZ_=y3(%S(%uanQWx~TLR zwRPNNXp6Al?WsIjxHYwNX#Z;My!!XN7w2HpO7bKMXpnpLtp44L8GIX(!LL1leXwh# zl*D;jqg#u0*0XsFm>N0=3oKgYiO(m$zd@UexD!8Cdn8Mj(=i~DI!-XSMyPi{3s^GS z@;l>>k01wo+!m4}Q(1iSlqY`)yD!fbFC`z?z`#~)5G)+R{ToYCtMdsWJ}UW=<-q_> zAZ3S_H!3C!MvtP!i7W5DG1#9ZUHgoN;^pZgNwIZpq*!-BF1$_rO%5+VpahKltdrwge7ee~m8S7>nV8p>5TiNek>B>jK zV^Y(k&THY{$L+f+HF9~Z6q7un7DA46?3`rV!1FZEw0mdshYNnO#+9Q4=BN#zPH z4{RU#CRG$Yjo*b_)eBqg_&5J%1HJAjr}*y!og!ReAiA>fEq*&^0$yp|EY3QYGuQ2%7cP-!ifL4J)daEk}tWE zqc~a;+etwM_wQxn0WF&3MtAySfXEWsp4FLbimomH;})p@*ZG?(>Gm%>;%AL6P=zG; zn{WI1*cj&Z_!m6e1~sd=0(`N!Z!pn+e2&6(J9;o({=D-=is!;e!@8&HeaBXq`r6S# zUZ+5o6Ys-sx-YvUBjast>?XeLLqpLa*U16mKFLobVQO9d*ZUT`&D24lQ`+)(->4UV zqDKV4Yh3@FFU;FDj4(nNyvGEV6iHzSOMlL#ylPuBZ9 zF%M=O*?Lu<@gnyqy|?9)X_0g1X;V4i!b<@j{o?Lt+sfQbM^|Ba8OgAe#Y#xC)mRw^ zm;7xNGpeODml+Qn(4Etbu}jZw+n!qxs*Q%ni?Vw}isizfQwD%&1Y`aGq5K! zc*78QWz-Q^jk2gmXKsF0g~g^tM`I-TC$I9xQKLIv&hSK<-Y7(e4TZL@6-%cr(X=`i z&$(lwSjCeyA>SC3u(hn79b4tEsUkZiKg4~8#M9{@NMlr8lEsLide z#pdl?Cb2gG;2(!I?eyF3P{SV>;`^^HzsI$2qBT;rQOD1LgVw5V77F3Vlp@%8KprD3 zVvrj<{p~Q({VMhu#^^@|4ODZy;=K^KzI~^|o*vW3WZyKPM0FS2f>n9f5+vT)qCy+T7PN0c=qOcz@ z`-{PM^XXJ1sZ@OEE@AGiU#2M)KSkY`P%j1P#JV;VTc*kO)D^)p%>vnb2ws@Q>khN3 z%uBW6oV(pa$D&=D=PV5Xmi)x_)5KN$DWYR3&{l%d+b;4W?U@VD)G=sdXtQEwEVch| z7^(9?rHPnFjla~4cx&2D`$l{RKf)QwvFn}(vEG-~KbC#`4IijS{d)r#eo#6?Zi@<3 zeSSF{a;$w`_}(z^o5=HUtRwQ(jI!0z)g$+JX6r@~GxNT++iR(p6~=_CT73L4HKq+e zcY|mfKPkrk$8-*wav@$LA)VMgS~B0@VL|@)z!XK3?b|`wDE=SrNPYB>!O>5=Y~A=ip|>M6DFz~aFbo!`*LRPs}y9b2@RlE>vn zaPOnJ>zgegj}m0_-eu)PF&k~f=;hnfL{TF^p;8OS;aPn$#l^K+$o+ZQU%z=#>fmj` zlPY(V|8rd}%n*}fu@lr0a(8gmrw@P-e$wUFcmDg-RnX&a+5%CE`)Cy9LEX?V&L@@F zk3arm<~!MxI24`d`&%gntTD+xTg;a|KLivV_<5T%)V-q|IO}>>h~Jw8m4bE&?I&IR z8`>H5A+Ubas`aM~JHe1fRq?x-seWFGs7dt|9c^83l0TFm`=WfH@CvZOHZ&asfm+;7 z1AD=X#zFaj^{Ay{9VixOh$ts`wN$-x-mg#qZ-(s?D z)~?>#SrknLEjIzKyxUR1Vsc6+!zWetOqVs(W})d*Hx1i0?v@*S^ugp0w^553Xm2p7 zJPpnQy)n*|?7U=|aX0NyHyv5@b~yW?q|Y7GVxr%mt$%Z$zt(v^udA|Ghk`pNzH>Kw z5M{AltGTW#^jb+8E!pXyoOjsip5GZ;dG%+>CaJ-Q(6$5KaKoOhWI~u_LI~0RzD7-4 zl{j3s;%%tiUJ$#Hau6M49>j+=;ft;f(tz zS3CMZ75j;5hNJ1@2}iQ%q}SnlFA|M!YrmX~H$OJsaPByqlNxaFoBNem!*6_fdf2dm zbGf?Kr$ige=@czSxkZrWzh<#m|6d&(ype}sW+!ewwvHCjZ)^#Wbp1CL$p8E%;9X$f z=p@`(F25VPp{R|+Km6Mdz=>Q&(Ukoopotl`%_6@ z;r|)L_g?d|oF&%n91qn5R?0Y|@Ls6VX)bI3;_+AA%@}tlas9tTSQQal41WjKUp2Cq znvWYLxm1Ial1O^RtTMBDW$yB38E%Bi|7%|RR)}`Ll8(^XE*~;Jh3#S*Fq+z63F_H8 zf7@dQG#h*yTdTFWrwZxu^n%Z<)2b((6MP73*xWMCCcOS}6!dBvXL_`DE7@Q~xYI)B zVFit?#GqLuTVK5q>lxkQBHzcgjJdD)0VZM)c*In>fnq8?R9FRu2sb_lIWkoP0|Pq- zVmvCRY{)0vqb-Oa&!E}l;bhi}{i}JN{J;Nls02IM7xC1zbK#eBxXd4Gte>Pc)%~B@ z{r_5LuJLS`ZF8E6+L+Ve<*kL-z`Jc>aTpjLfP8(^mjnYR3@Tg*_`Qp3;$edM?c;CT-|L4Bcz!}o-?w}x?NtB;3M52xK?S3CJu z#I|})wT7N#-`Prm65IP3R9WoT+A;S`gch>cW2(yY|NS$B|DFl=wXjOj#3b|vbvR_N8@^J|q$E4cx?t5x`I@F*p9!+pl(>hU$AbU8fCzuv*Mp}I+P83`@s&U~ob8io+C!e7dIQ3< zRd;hof#*dcz$hEeiP!df`p1p$9iU7)l^g408r_WcTxAz`6^qVCR_7+MT z+M%2i;7=qXQCUEwks&h1uDB<-TG}z_GCCll#$PD~x7Yvq8h9Qq4IPskTxGC!RvUuR23G>3~{Q=_wAxS7yo-n z5H~A;alUW;_aE*k_ihveF=2dXh;V`4O~ZO0W+@0W9ic6Pq}xORkmLP6!J9%J^Smao z(}CgcMrB5Bw|9@2dV%TFdCG(RMAie@1VK}8@7($qR~hAv_XeY82P26iL-ykwnz4X;tisSyutxSqFIJvbG(uP$ih~qh4ZghVOM9c_&i_(4pdMO z6&QF$#|V4b78D@A8GARNeDEN$73H;``;U8)4ITh?ZX7%Mxrp;E+T)IL6l#>v{lfgg z*-z!X>9~TN;4@>k%PfVwgE97h=A<+E-$lpBSum84X8iS$Aym1Vc3dk>2(^w`U4j(d z`I#2eZnw#Y6X z8Pr}wujn!RM67yVLGpE~CZm??{ex^q_xGq!4SZY3`vgZ7kE8XwT4$*CdH#4mLqGGt zc8JZX%c+0;b)hM2K>(KWE<>ui9?*L;t9x~84C7e_&yE4VAl)Va;we>mz6FUbLR4aJ zUO4CdP7U9m$7IJCyt^?L_Q)da1LmxqxBNmWP{^SAF4>oS#w&Fb+&%c^p?x7ttWB-C zO^qnu{fMPN%Gf=V?_iEqGogOm56cod5(%WF6T0_T-ysdsf>0df@e8V2+Aj;l zV-|J8_-ax+ZakH_*2HTeEvWJntCfBpZf`lk;fK~gqWTaG*1SPyz%>c2P?IXI1Zy92 z-jaqedS4n|wvgW7M-;U_g4aUI*QB@3b27u&$^{$5ov!z7Ki90Z+KKb{Fn*TjbQA`G zSYzl$GpK-xf}`GI9`0>?HXv#he?D~na&PGF#vsD}xHgMez(G@IyvGy3jo;%Y!W1To z|J_W(A9Q;T>KZdcs_P&G0HJ62`;;{AhkZqCp1XD{GokVODemt@AhW1YhU=|yd5<3Z z0eI-u8{3n1Y6(w*K|so8XD>7T)v(ws;J|x?J^~yy2hTzzgQzgSS9; z%69g&chE^>mE@GSN;Qdlr5flHRs8#&r6NW()KiqWv+wV1S+iVb2vN6ja072JB zjyK0j)ftKkb=ZvuN0j@^T9FEbO=pF(+KfKJ16Qca2J`dAOpJ`wdZLn&qHtyVkeo6l zt+qzw%u#P9l&MYb1_uUpodv$P5an)pEyR@Pca%1n3h3vbl#j16xYHWk0G+qYroo)s zi$gp(Q*pxLR*{<=c4mS?=PHYAtW5tyiV8)E&yo^+yZhY-7qQ6^^pf3+d4n8tb74Qr zd7v6tFjWYUVV74iKqwBLmsJwcic-M)Eigq5E&n44wD}bhCNLHOV;LSD!2WD};-4h|jo5@L zIs1D9&08!gST7X4jnxns2r2ya)b}+CHZ76-+-S;f$OwL(m3VsbR;&jcd}Wh!bf@_Th6O#F9d9{sMy*=h|o)(P-7H$1Rz$?m3(&BZ3KBVkNw_l zjt5mKw~Nt*hpWd0+8*CrI6v}eAWw^=`p%8FeR#WA@Gm8#R;0?+ajn&!GS3vd{aWO& z=%4oAASyPyEMy=B5_cP114fc@A)|THLoBghp}@cmDhw{_@PEn-rwh(G9}*P+y|AFX zt8l%(>Vsv7VPpjIfV6{WjV1CURd@0lfNxyEEw~COm z2bPE{3)&tF5R5$ZYZ9d}R!GmPhx81yNEYGPJwqY=t?Ou$n^+i;k3rL^S zx*ket5?ta^vo9{miqGGH1z6$5LN$n?LRm~U8 z6D0v+f)B?Q|HAg6Ak#HT5CIBUZ`^uKp%~^u4)}TYCj}lJk09Pbm!S|xYCnnFNosGs zs*Mg)dPAKiu9D4z2AjmiDydEfS33l&aa6dx^~BmGKByYOIfw4q9XZ7~C6eAA7!NOVN^$ zpkATqp4wfw1=#y{^s|x(gkXBAa)dD`@v6d7-F$p=f-0zUsxMG~GUpHqJ!X*j=+_wx z`09VqY%}E2x{CCh0M-rMZWC!p#CX@x-o`?EkS0GHgq0EXN-0Ts-SjEQTD*r`MY^~Z zqy7Q!0a-x1O|hp?W$%$xkkoa5n2|*N^>}Stt3$a}UaNJy;KFtn6rS_yCsNcoi3~$- z`sZ^u-Btz{sA)Brh}k1DFMGjd-nWO7)xvQXu+}{d_(l$v9|K)GrdWYxiU$amHEx1= zOK#jA7_j=HzqY*H(g!mYKjYo*KhG_{`+h#{>YQ~Y$Q@0W z29#`6g+8i$zL=Q2y-kS6{)O~nm9$401qGGaVTMMBzBqIc zAn2C>6}|(WEY@VrX+ksKIS&&ADE*4jT0T$egFlmW4U`V`jKZ0Pgh+)XkaVFeI*6ps zdHzJ<<@-eIr2qs8l7-$mf4)z`F`J?kLMzrx@k|dpTyyssdpm}#e!VwWcRovlScAQU z*AOC6aO-JR%6^c0Qq5|US51^$DM356-9Vt*EphJ|xANkEJnnsWsm-7MnM~|4u@la^ zge6^8i#Do`#}IQC+vH&s@L`j2GMier{&qUYW^=!Za(r35h*Kp8fmWh=fuB!FSAU|~ zjhB_#9aN_Lq`tmA9On}EM9d@5M%2$!BD5rI8q)Ng`tudKO2c7}j0^(<%*ifheKHz@ zjU`4utbfvm{Fd5ia^UlO0;Hnm8}AG%AOXv6v=a)(>9;esy`0A$Y2&Cr3+WQcS`grm z$NLR)ZSE`IqM^1mO(D?7%}Gu6IP#nP1n@Rwp3c5ixgei221xiF;)Zz;A(eChf`C{& zN7=3z{h(-SlI*!a0hY+ zdWBv^$Sk#Ni)t9sfNpNsQ7tgh&OH(&Sxw616;n}O<0}7A@k}NH#2Grmy$Wjzo~y}+ zp^f&hVq`kM*ohJYcr)w}J*M0n`KnSxcE3XZ?aMpBaUZ!&!rc(Vl(u$>iqkuUu=8j3 zahcZjs`8X&)*rgWmvD5PiD_sdSW!uDQiLVLc#c0@tA3p>|>5K z!vwxF02;&n;9KMrM+Sz0M)z5%k_yYvzH{^>(m@GrKK8V!i0zjw_uiCC$xhXkM!Uan zZT<{K^#gSR)9@02zB<-6{t@I2Vts$Y%aD%SL0ezOiMPw3Z8k^vMvvzMP+d}QCG;rO zUPeC8?2HnoWKM@R*LSazq4(KUD`NwWOq7(}5494H%Qm4Tae;N;O>izk8QGhSC=b&~ z*g;wozj(jhQ1T&sn7`V#f|Y`L7ERGfjGOH2yrWbm&4c}vu@WZFDf7-`?iOw++4A|u zfD65^FScikHIx_~&6?xl1L?v}L<5jKL@WkxM%|^`sn2)|zyEdI;5uL)k`_WtM^|9; zX?pctzJL8C?4IC$%&bBAF_fGIF=#_IeA8ut1xlcbKuEj02M{xY9@=)L_ferGp z#Ex?0Q|Jx(BW@+`3Ug`BE7^UE zT|bRFSLRfPK!z{=1V&=VPfDAWMteEnSol(s)eJrY??p3w(?kJH7Nj`5K|%Lr zumxmIObhfrwEFg0V<~SV-%{i_4aeu;Wh^$C%_oF*tXsR)#drnv0LoA;8j>4AMEujz ziY{giqj$Q-Cf+8QWcY6d2#2g3ajO-+E2y#5IF)axdM-7S%R{~RdU)%V2_LD0%>1i&FwK_ZpEXYI z%c($wj#Z(n4+in!Slg~(S$=)y6OcV8-;F5r=!F92oymW|?~pls)M1k&T*nQjZgUy! zG9no1c9>l%#@=*D*D35{bSHcdUh}4#Mk^Cl5tI|v9o+OzJd|=<+40M`1yIkjR@Nis z%)s;yD5yfsYZLl)KIFoV(94ywvxlFg3~a5Hqnt0FVgQKey1Gm!YC=OP+|Vq(VXq1to{yb9b|gDufWDB*m&|KE8HH8_Ud za+PNh7Kmg@p{f*G1!+tzpyXZ2aZ@Gww_@xN$3N&*J34XRO0T`FMB z2(|CLVRt%J)!JNdwEaU&_3~mA_;S(qVE&c2XI>VrZWOFd7@|*uAr7ak?0{UU-OnDw z`m%SDf4(>8hD0P@P5QvR4S*~%irbp~G;lGvMKR>a>kv4f2R;Ko?0~cv0hVxw^XID~ z_&<4&hC#^CJuFjq#`?E!59m{G@;9u_z~A;Bn`1Afy2*>BSL>zrx@A#ChPjmnSs_IesaZ|RG zv-@h};j`>e(kPBy+87`?@5{5Y)cMm7c|;d7{GpJrc_VPioHJnx4OuLL27gF3s3x<; zmY$ogbK9mww@itszK$-EgM|<`IKA0v!Q^l%gG*amo3MLCl`_%+u^H>TkK8Tt(R0G| zORV}Y6D5zN9X>t=z?wchL-cbk2(654`@KZuIT0qOskt>nHToAhGLSzPZXZxTi13vz8pnFtwzrgDPe*eo5wiR=z<+l$kd|6cn{ zPudA&=aLBzpI++g^bnMA^y6Y`{}0+Sb4RaIojAY1PiIdIt3k;=mw&Njdv!VU_1p9u z57ylK;*;W6Z_oJpRHCEBKU}P~v|M!gUHtSV{3xJpgn)D)@TpO%FZcv9BIJDW1(XH< zHQ`URg?^k8xZ&vbP|N;L8)8*)tzjaADO z!3rPWt-Sw;q>P==-qi~)%bTca+jtZ91sAP*t$#B5K-5U2=k-xstK4;=x5x`YW);6| zm*e@=3wfr_5Ob}RS5O>3eOL7^z}4Y=n832*|5*UceUQ>Dt)-fNr?|h>+H#}c;BrB- zF%oY@w6yxjRARGBRL5~crgRz5JAI zbXl<4!09ea{DGp;fd`ZTL%m`@HfsJm_HRVb_(Q~3;felP43#4#ac`8uE|+#gaWpXy zI~munG%HFB_io77H%etFdtCRL8)tQC)>>#hmc3?-MU1}01s{LiVW1a^(>K%Nh^6qF zzp=R#G>Y>XNe@I0hZO1nga`--3>7woDbx1*j1nrA%^P&Euq8MNcJTqfQddj{-X1D# z=V1tCl{UwrZImNHi4lQMSw?1N`8Nbq+-`Kb9F zbQ-K{dSVF3P|WjxeKf(cn@-oZ83x1=>_QeafvN|bDGV+^4n%0+XCkfmMA+?f|PJS}y z;PMy{3ND&Et(&wl-`8&5Yl7t!&~iy%g$Q8TB(yKq>g)Plt=3FrCIMmcFFhJHtCg5M z*i!kROEKS1a>%{7-gfpKCx=rZcW59q=z0x12afV* ztT~7aa|d`N+d8fx)GSr?EvQ0z2bx9YyHw$w(+ye7nyLyfwM{pXUhIU5-{cTU1gY* zO@24tFVVZhinVfGt@fWV7m^d$X4mKc>Ik%=D9IQ!r66ddUQQWFw}{ef`JnMP7`I_> z_dbT(3jbHpN6|w-U~y6R?L-yjURys4H8*ycx8;;vO#?5lWhc#noF0{+|D}Y!fZe!+ z{!y_AyHPnfQ^GC|Y~`}~W$dETg(WdtZZ{vgo%!^4L313q02|m6s?h6bG(qY?VWD{) z+QRi0@DbhXmSU&Zm*}l4u)%XD?SnBC4uFc+Ygv5wUn5B`0l*I0kxcO{%(FcA-amZ#l z13iiY3Ua?0OyTZN4?|+(UjD0Ia+kfH7JwRuze~K!2>%0+QgtRZONZ0PP)JFaQAt%T zI3!=^eGOQO(+Y2AXB9a>ZP&owVwcsrq`3haf``*$uix#MWbF?kXcTEY=9LOv{#R}Q z!%Py%H&uNFGN^)X=6gV<>m$}=gL-Q#IJI4x%#Xee;fH=<1=<$!AfzjQyleC0OE_Zj zjI@^;N4Q&&^VKpsB#0UPA+xzl_8C|NL7-psSkQUkC;v)};C~9z{y{N@C9fpXFyZoh z0NU@w@G?8t!%d%G?*AAseezd*R1L6>t-jCKH=`p_Spa4c3*IZs%ZDqME5IdWsmJ0( zY}_fDEt6D6WB2`#v2IUWpyFF2@a40>r`PH8LFue9#ZK+|_(CVKtInUb5{Ij=xZYD3 zY}g`p-A{oFwx4pnht93N4xACAzzef-8ndXc;8O6cZRwl7|NM3q<{3;MT?mnsWZBoW zE<0b`RKEWZ^x zZ5f@N!l<3cBbf?hEAps1wHrc_SCl9PM24yY$4Y{S-fQS5y#!787_a$$*vr}{(aUYV z^gbj%<(A}9VPS=iJ{HU_`?G)1qEax|)n;wjU{i@|T8i27i5i0^osj~Kbpk3ezR=9> zRT)@U`6sRTm%bs5_m|+35oD5@k8w7Jp2?%?^!Xvi8+&MA3|Udnod4KGGBU&8Tz+Wab|Tc8ZIVPb5$YIsW?p5Bke$B~idg>k@|Q z**DIp>iBTaHW41~TUk*b)iZ01^pi%GZA$3T4OPJS@r>68Q5(uP^rt@8834LFuFZez72NMn(yrXBs^aFO(#`>#p;5;)e{FjWjXd7*+9_e5j zpQT2M#d`T9?Ex;Qvclg?edl78f1ImIH1KA+?aPPxk|H09&qhWf0ekqy-#Q=VL=Ebt zkprtIZ9M*6hqkg|RHLwqla;Zt=8hsSbxLt?Ivrwwfhu`6mh#dZ zf!Iq2bFBi~TVsFH64ZUrQu)+;(i96q5Qm4JjDGtE1}f|kDzZx9jgTg_h+Tw%iny{m z5D8m<6tH8a;GC4G)lW9)Ii0Cri8=Wu%m;BH{_i>NGE_vVXZ?YN-A?tSHpFCEui8wR zN3Bp;Q4X2k(rIENnRNLsqEpJuR|~lvA0;TOy{M0YRJ%E!{uLf!?alLCiABnxhonm= zzqy+jlK9TVyv1dOcvdL#V3oHl6YaVpPc4d?^Oj|e7*iEvUH}u@B(Vcgw3hv0th&<{ zWhy=c)IKf>-Vx5>chis=u+s_WoNm2X$KQbj868)DWfU-Z8}_xX_nDI$Pvw+yFT_Yf zNDWdaJpk1TTKoaGpVi^Qn4!?tW+8JckWF9v&oBxVbjexrJ?d`PZR%nG67i1UTUcDdlQ{Y&H`MZgUC!## zz$X`q>MK}ERzE6YCmBs_SImjyECcaE$lI;IeX5%MHtna=B5d~CLlk!M504>T^a-lV?Vh`tuD>P`-fwxb&c!HqtNEGDsMtXj^ab>u zfMcCXJ&p>Tk_GdAJIMgw)xS=aE^gs1g51sRh>VI35yD^WYm%h3qb zJQD5<90QUZ0q%xKSnn?LAu|tAU*P}!>)aOU=l%Dh`RV!{5cLMZ-Q048g^v)V6cs5Y zMWDQp>5{r0Ks9fvJ_)RK$bO{-b&q=vnQFynIph!!_k5HBEWA7iCFX!1}}gzU$g-I0l*hmWtH7IfyQKFCnP(742f!4>mB(GUc~ zH=Xp^($ixAbhYtxZ9!0#$9&I6NB71*eOdmi{8E?05QwA~-AH(=%vA*^zzt2e7)VQ60&sN$584I6NlfV{enB@j5Wwt z1?*Pr&5ywyazH=(;Ki;mRTa=dO?*9!GWURQG|w6KF!1!hxJJ?)`_Uo}6y~2&1A6v# z^c$kGKCvC7k|$v=FN5xs<$F81o#HtBq~RG`%N-xj$UDdQF7+u$Q4>GT^}A z=wO|Lg;jWZjs-H>KpMifRm-UVVu7|3~?5m*UDy(gXH_u#Eiq+lx} zPVnHwiJWFw7S)i+2a*!3e^zG@FAhMutecu+k90`KOZ*liuDz15q*4+fEWWI5Cju1CJ*eTfFs1<~!9S(%RZ0Y; zoYDIfZKu2WYZPrXGv_Pd`kw$I$WYK-*%F7ysH9Sj2%JVKsu#+L|CZpQm+H;^jYh|| z5pVwF8?RSoRc-wtiQ92O*{EcAjM`*WnqGrJxW?2d&54>@mi{{sH%u@pN$q?AOmd?Y zWTzyah`{ii$;P~)uu}`2eMrB*La_??g6Ri@9NVxH@9vzI6C%p;dMjGp0g;>}_A^il z6qpg_v;Qf6;3Go~9(yqcmEM1v!pc#V zm^D1Fo){CMzPDGP?L2{!6u?Pmj0*OAArjgygdtM$&69m~(OXjD^$K-6G}MJL+I510 zD~=w2w54L8XNgCqQ0YJ96QwrO_dnS73TTQ@65N8kj=oY8r(4fO>Hc6G$^U4CSMaRO z5qAgOt{BqV{9_&7V5X5RRuM;MqW$MHQ3|h~l-7Ve#izG|=oj@+&GZndGKaFRV~r5b z5vi{!nAe?5sQ3Sg*=-(_e4eto9VogX(Y(w47X}+9~vurD=WtO3ufn zO#US!aOZUX_~8fjZ%`nEtwp;Bay1!>9+rx*Ne~I}5|UJ(NF!c7Fg8v=GgQ|-k9v9c zMc?07@0Ls&@V8PQXOq>aYeyI*vWTH)--sUWc~YQ&8UHMQT04$hvffp(pNB-8wT+!|WKISif*|27>1C@(pTJhGCJzVA) zDHbzMx7JaY1%6F*q?BULk1`wkpxZ`jxc#(Ss-+UF5PJ-ANAU@u^FhN|ba(KUO>?BG z{+SfPV66DBu!8rE$~5XP4j?>XQG^2KTR(^dU-Y*WEAb` zj%pEB)dSc$hN zpl1579qr-P!ZuSPa0X)^sA@Pq9w4%GTKdyG#@tb1FcjeIN7lcxP(>#t`p$`>r^YB@&dp#~$t|{}EYJ9`#hy87NhG0-&1;HnU(){Sn!=Kl_ zoPWyJwU+%-cLex|Mp+H|lYZI|7?S__tb`_7Fg zp-TsGBnv-#EOn}YZ2+x#5Z~NkT*-@5{qftVg@);rut-U*I-Wd?DB6 zj8l0{LtEXVCYi#{iFj0YzTO7$_B_{g`g#^21ywYmzO5D1JO}AW+sLmks(#g=W|a3H zzvtY*8UMVp47Nik)UeEzDl3C9o z|2weqjb*cKgVfM8yA{!h#Q*g&vh)CaCUvb>8tx8;BowLnjc$`Vskc@d;v1JBsPx;xRUT<=P_;*)~lg&S+kjC=<1ZLhpZp%cJ$c^F4 z8p^D*xUtepI2>nohm(2{|I0XL^t5`0TDCj>cvjgAGoooG->!XJH$RIXJUUXAGQH?X$j_|p1e8ir2h9v`cBMx0>pXwmI|FX}HsLfcAkmW9@2k#~MBL`s>+Mx1{#a{!FV zr0`R55I`!9ozyPh(j?)0t&IZa89|Tsb&`LbiIy@y=jQ#C#y%PuQe9+y4gwS%2r`P3 zhz{Ud+p%_jXNJ`B4_1a#FWE1S5*Mq!?3;!muAnKKpCYa)#ct)r0rTl;#*?)gt@?=A zE?MpLKh)daPBJUu)BZ3gtVhTIRH!e7Sb}LHqft0-q&FH$-1bXjEle6$I8OXoY(80e* zb9u6+oI>fxnF-jOml8=E9jY}FhdkMA`do(p_=f~=WosvZ(FCHGl)7Qp97tvJpZFrl z74n()x^)B^BRhNI65{0ZLeuig_b6?lRHKf+Q>bsuuM6`~vHM)=KQYYCl@J{@gdz=R zt~9A`xFb5jKbzp9siH7Z|Nj(;qHYf+QXi(^V+e=ZDU7O5tcZb7 z@~5_rmj6um=7V%_k1`zTaQ;q|36#ks!;v$cD~nE|L7zR;_curglnXz=_^&{cjs(Af{$iVAMo{;5iYN=NN@HsPJ~qy zUfUlqnVfXPCm2-m3{V}u`49DA{z~`Yd0!D|Hog=cQ8lb0<2kTZ*Abw z6B-NcZGJzA&LXnMkNyVj6hR{Czv&Kv2Y7@DrjQ+G=-!8cZb*kKtF4WBT) zb|})J(=`w7ir7uj&nMipe5jA2tYb(gpgdz?2GyJRYlUTyTm2feHF)sH{WII?ZMR^jY^m)?|w!6K|$Rtn@WT>WR4EPgIo1QRt27^SBfMWp>8 zM^R-@`6jJr8GkPtFJ6%!E`Ip)MthOHXu4%_LN{9LP2yc;bqAqF{&bo+^z~)8m}s|# z+oW0T$ibw+OZQtuD&2w#NLS>*Qq{1;v@m}EAbG=H@$EA(P<=C7-j}uWI<>)J+)ezt z<4{rlSO1F*w=1VaiL|JD& z7>Tv27!R~LNEr#H@>i}i>M+~N;t9cUqWbS+s87`3_5cRmFTP$%MdYYTQS`Bk! z>~7fxR;im|5=2l)zW3!zdSq1e`0x$MKiX=!T>0x*IB>zCzwybwMK33`QF-SK6+7;i z`gg}}xJ%d>|0w%@rgA0`0jDw;ra4mK4?Sa|dv-S0Fc5V?$&a@ftg@`5TmPQ8bTsoO zuUU|XA521Cm;4A^i#sqQm_cp-Z+spO3%*L>w*jRQ4Mh(L^c+y_%z_=>n8|HcZz))Q z8K7i3B0yw$9bPO=tMg+J?C$_S-vLmws)KK!gyS`#(^hTH343AosO^g$K-uR z3)({-4?m)>7%Dt=Q zaT|f4ea9BZkD=bELkaLZDp10}^<|fPReZUUm=sDC8VTuIpvnIM4`aZd{S2K6l-RV~!HkevonSVI?F zQOWd3_dSz`N76xcsSx+M@Bb4SAsPJD03Fk!7qed^ZU$?gU}9p%lpC>IBa=0JELnCc zEYCwfSNUNtSIo(&E?x43uXNS(boG?Q{PHMyNsFS<$WjG!lh@;1Vg2rPn1tl|uwFh5jE6uYW z_p7P1ichIq7I1~Zt$pYef1R@QX-_gq#tqp5|&!gUd>ZCyg{-{iuG@ch|>oj~x zc~Q@DjAOJ9^58B*ggIl8v>?vBpfD*^`tlx%J(&Z=V=zf2Kx(7rkEDk^rVktu;<&Bs zU;b6kh~{7LjxP<~OmyIu%fb!2*?|{-ofYx6B%oqMB`Uj-a+$!z0SJFEO~W-6JCv`5 zm)6at)Wb3^P0a4YkJNf^gbgz%^CAcT*DL*d^sDiWH23vTH3%lv$Qx~SpUy8UBdM#@ zUU|0J(#0uH1NRhMk4BZ6PftkL`K5QaH9*h?6aKVJnR>UN2dSPQ1itd2fDkN%i94lO z@>$3hrCYPp$yF{@1dG zOJ{)K!^=m)TZb%~x`;P;FK75+U5Dp`V0hb(H^uq~G<2!2=7Xmbae%LxjQ8$eOmW)v z9qs*eB3KgzQl-;IK;s{g|1|6|)d$FY*W#3gCgk?tYsQ;D`B?r!xg0tFS+s0+`7GCF zkE^H|U*y3d0RJtV>Mb%Vv837tgS9OIh~ zPW-RkWfS4A%m2wSpzz3BeD2|^p27bu{u=aFC91w)BKr?n)GUDYCl$=@SZNuHzJie9 zRXy-YzOnhJ3-by+MOeLo-=xS>{7340{qypIKw8aTvb|+7%8X!s!s0EzU{lP-v~rf; zq(+hQ3JMLAAI`svin(Oeyt4FK2pk2?+P_>irgz|bFSHhc#VeulJIU|eredfX3a|5Y z1v?Ufg5}Zz$_?+MT#SPwMgk@tJw)OW$>_UfTu2g|OIG)kj?kPiy_)D{*1L-ZNo-Pj z&;C68D&*(EGUDhsLe|#PZU2)^_0sp*taADHA8|Q=(DyZ_6#|iiZI|X<$Uh_U)Oh|Q zAG~mvVt1li(-7Yu%JW3ec;f%XZA@Qmm>wDbmn<%YVEgM?5&}U3FpC#BOtOvx07Edo zfwN$p*~2BE>9_Cnw}1BvNG(F{H3Pts1tM=TRKMd<3!i}fBVPWwnRhkfP()giU-|$* z-N_->1JRh>Un2CP-{FwK&c#zi_sM1i(|rE-6sYd?MULftjswEkBtzuH0YoW56RFST z+n?bZ{VmgYJx*_h&ibk;xA^$#puhOn;ADp4Vz>oHQV)PNvYg!;)u#y(!LpqRN8|HWtfxjB)Xc zFCvBPN|&41`e1`gQr{@jItSo$3REqUOBAKhuVsC5uOsPemc^~sDu z5e!;d^9`aBE%cQs1c@;gozBV>wv1o#*rn@W2_W zW&e-n&$qYjN($&HF&Ob};J42q& z3RGIb{a33cStgHK9^!uIA}tdIXD#5tqvhztzP%Z6|5Fn1jHAxzYlqSEn#SHo`>{Th zUWQ`7X*kI<^k%DTPITO+G$LcxzW%jk%6w#Nh8k|33wPK?HJsE4DsAo>en^WmsnrZ@ zc;B=9s<5!Ba0|!qAWFx$v{D^Zsk16lJk|Nm>!*)J^A%3rkZnR2pIU`IVyA3$!Jp^( zpSZ;Bf9s@zUy}W>fM4NP9cOOey5!rUaC|lLB7z4gika63Q@JhF%wPN*;q^&UN3+1% zD?BSZLzL?ifgf3F@jLG9i|HO`?nR2<4itZ}z#jl}Tf;gsIg0g<`QrCV^)D~mu+P{_ zK=@Vf3*80DhNU;&F0}6`H<*<5hRtQv7P}2b{5e8RK!5b<z3L{Yf(N0l|BC+N}&Z!6RdaJ zJ`2*J+Q@+;BIrcw2@J?Gw*B*vYs?$}4(*_YH|5D@qdn*N$4bSVtO3b(dD%x3WB-IY zGg;g#74*a?B;$FK7%lwM9}bLq+5g|q32l7%2owKmgoztE{V^RMMmG%p9S){RIO><4 zNeJ96LHi1zfEA;fV}QrLA-Cz$FALc`N1B2t<#uqv+^DTbsp;Qn}iM;|rW zV-QM%BW1nunyM_6&mT&K6JvA45fw_Y*XFTFr_f!n&rv`6F$9%c5oL!H4WuL}$%+C~ z!UrsZ*`0(IwUJ<=X!%wN1uokPd@LwX4=@SrOZT-d(RK+=|B)Fs8l7+pB}t?+0Fr)G zrkSI-o|zn@@AIzwGk!x}qT@Q_7hJpHOnI(~;!HRl+}V)N4NBz!;~efN5Loy#*Sh_K z>ljVALy!k`H>CJOhS_6bFQ-|LECmJ-kvrbZ6Y<4vP*s~^uQT=^JaiRvhdILz1{_FQ z(Q#PzJ$_vG>?XEjBXtCQ14N@~@{~tLUiG6{c_PE$Wjk9-mr}!P99(T1epyhFAR23H z?w{j*iK*l5WQlfhUcJVNpSzyebgc|)@`#Vya>E~f(QW#2^ z)BY^zBUwtEtdcP>e$^G#@95M#AuuD@l&vxQP04-*G=n#Wr zHL3U(f&1$I*Tm(n&dm?eyY4vBtYM7pvCeltTu>hP- zMP@^jNkdi%2%a;Wl&}C`<=H_{nbjm0G}uF!2$jCO4io6gh^Jf=K zMKmbR-n$;ip%L3}P;?nj688h%-#bTXQkWE6Q$?J3t3@12WwguvLj>nMMw&n+O=n zT9Cz!9gy|T75mAgMvrJ@V~e!_o_gu_$}r8B0w+cVJ2@VF4oK~A-o5Q#y4*v)wJy6Bh2Me=1a=)%ry?y6g z&ZT* z{C**4o)`wMCFYcS1wx;09P&OOvf6BKVS1%Ea(CUNgYe+T%aqZ!bw&`BUoJj3iXy|k z??q#kUTeETm(3Om_7rg6?A{LR{P2wNz<|~ZGHS_Af)!tvs#Du7*IQ}1WdGO*emutT z>%yS>YaTq{JN*w)jjLC^scs%nf@=L0=Nmx=aiM&83O=L7o0xew0UVsAH$d@6b!r5= z_w%kviUa58-DSGMnv>c*sN9Rb7i0d z=XqZrp19i$BF->KcWeJ3>ZF~fhXKf76ZSAf>!QiDx!rv94_^Q#luptj&CnunM2?Rp zge{bTiUx+y&uQ7AVNh+P9rPGP+r6RKPLe#XGfjpU4tl;?at36Ou;15dPXwO&bRxTKf&PU!GSS8b3XJvNctCKJE*=vo9cC z&1-epV8(NEqn|Y3O>^&iH85#&+Uuj02SvljliBRp7R~BvM6Qpz0sJBDOF7%4o++VV zjJp{KI1eL>Dh+!}b5!J-fEIrD>``(3`r88edNPFv!wPO28z+1H4=<}Yi3VQzUGAa>GpVaatHVN`DLcc>U$#>7 zKs0!TgT*g$|JIS2g$RVH!8@bvV=dyND&sOUH)HGj!%|mCzzWyM#+A*$^zM+l(2V z+>O=O{^R*^xwEYspfIq+lqAJXV1wZmtt58*Fht<5QxUC*IU zdk^_;bVAtdP<8!|8@nqiu(Y{ZuxGdoQH4^Yvvv zIqBdi7U&Fr-j-UD299?_qDP#L>k9;|7;9z~>G8L0<|$l#3~|V6n*#hrAj zZpW4K9Orz#S%zo1UT;GAcd_`tizU?*ARTm4bi&+++lF!K8LbMZ?DvI-h)^rDLhet- zw0OA}jEeNmY5Lz_UAX^#))c!DM@T$k1yEi~ZjX{Rff1&?V=dDKF$5p4ajW}s=+2n) z`Xw1(4tMW-L#m4`K+h{<^oolDsep!JD1&&Lc z7yjXKbJS)a5gHm?ZPQ9N>@4Jg{V-3`xQz>C4Keb>gVMN78MAJP>TwAOv6t9A%v-!t zsx4YfqV&Y+BhA~~a=6WV)mr8HYsI^{RV2<=TU3GcesV^g!Lx40iC$BteQq_2K)>_! zYLga~#adIQ6RSY23LO^Fz}-iYg+ar zab#Zm%i1fQzVR+w9p?D763p=we2Y2T4o+h+`yPz1ko_%dVB|0`5vFV)$KD!r@%FDM3qQ5Z zUt+ni-J|6j8v(a<5s8qdUAw|Gno*5wguk#5#!!L^@F88#k@QzQX%B&Ti3UzQ`?IOm z4~blcNoV`4A`d(u+*#M(*XsfZLD)EHGVSeOB#$FOS@dWOxQ|s{DHEOIx!|ky7G3{I zho-K*G+uK~aX%Ns=L@6_c6o~?dpti)@?U3leHJ^!@L$U{Ig-JCA+vcyGncwzFz2P} z7WTJ^n-S*cTyxRWrovT7c^BV%Ly(z3Ibxs50m;-XhPBy#!Yw=gk2xW2M3tX}*jat1GvDR~bb5uJG4f*_c*TUlHnY_yQ15M^XY(7F9o zg89)m>URdP;ENQ)D3rxQ1#BC)>|y*Ag?cxU%8Q1GETVnLi9?=@rGd)7vnofb$FvEt zULT^Z5!F-<9vWzb)~;uQh`&ym8jI_dGf#zXiGVlKbzoBX^4peA<#w2Bd+(06nr5Zg zOpzCc89uNAK0|+myuA8*kv2;Ofb$daQYc{Zu%hiR52yNr`G12O$iD+uE@dz)7i%o5 z?!JSsfF2)?P>v*z?)=aq5nrM4w@c$BY(6U4B41=*df$Y!=+5Mu?~XqFd@>Wm&$WGVNp}82x2?Xp85^h+UyKNh zRAwgtY1nS@0;8a5V)~y%L+b{ZG8IzFfw|re zJB2LuvZCjI5@#)E2wEF_Ck;e{&SMfM52pi0xlIdREgM3j5`k+>IWZ)6J6g!SBRx2+ax{O0qex!t1_EJ60Qb?e z*P$oqjW8wdpc&-kitlWO4@-)RZQv#8Q5)E=?rv@gS=SHBR7&Y^YU;*Oy_O$<6YnS~ zD4e``(>b%6lyu@p;P=yyKmz&Tx8k~;LL^}vTM?-;~IGNnzW)EcyeH9`?p+s zc{)>}t_~B{H+*A=CijI#llnBP0h^OJ5st^TTxbuS#(Xgq*sEjG_%Z?aB<(I4ji;>6{|7Ze%D&pvY1-2SWGG$5w$7K3!=6m;r@fYkW&!_8;5+^bFFA4DH zkcH*{e*!-E$G0@Fdyb{_++F>fo`1GuL5E9T|2Tilvf#C6rI zwFe@VM;8El)$2vpB*rd z|NZ<&ws!ol1Ioe|n180d$oT*3`5y~|e|0e9#!)sahFdoR<8Mp&9{nrTq_M4G)V5f_g;DFrI+L`m_kzs z`2=Gb#$^7r3=slc#Kow_m@ z{cYN$nWTiPr_Z3VfB%E`)z4yMSZ#ds+qaVKnym6EH1U1#0y`(W1rj=38&du;Ihtb^<#sfB#K#fa?z`G1+o3Rv zE{}d1^Amg!*MM?k6NG1rq<;Md_^Th_0nmd=6XMgpeS7H(IQ<8FC}T#CRb|KAIddht zV6@QW{7DEN2M-zy%a5&6yjU>=6|A8&rAhhQ{oa=H<;tsKg=5ITlJC6puDt!u+tTvR zJ0xwIbPyUY0Bw@X=`*M0L0BTZ{TAv}D5nBtz7>!9ag@HZ{PN4s^4hDfOYS_mA)Gr7 zyxx&(L8DoJ{e?RHqQUwCt#ZVmd!jAC0d-utdDN`g(>NKjvaLCP^IZJnrSA)6x zj6Y_|00=+l&tHHSwoKrOamp+G;jV!6&jDffoH=taW<~WQ_Nga3YdI>b<}aKtb!*oF zoffI99V%hQLkacbi!b4?yYj&YAABHHt5n6#mKoBoUq9sm_Cu$b&N?&cjvqU!G--iA z?b)(slZii1fKp>Oe2!kP%8E=7wvPi(w!HH$`LO?oQo2+b2;)~uJItF?5bT~keMV;f zIaAGO9smD?59uy1bWwga%#loYu9FwvLSH@_FhD9*gaQWRNSa-Z9XM3Fv~AlKiiuTP zHaQsY`yal?oR}uns@*89?rp7P+z$LVLeW4aATwspnx#s)Jh^jAg9bOt@8DnNPn(iV z9_{$328JI!`X`-}^nrSF!<%*fF;Mh3-;RWkKBJ5nIReU|%KG9aAA$>Gjk*O^K@=`?Tw^xGNS)fX z)c-ey($l6*mj@quP`>#53svkLIf6D|O;xN|Q5Bw30DaqS&1LG8Y07`Iz{8!M=p=Jz z&yg+Lw#i*B?*fk&Mla7{SJ6oHGdl$Hzv@n-rK+_@vfZm_7YDYo;o3nl){P91)H`l~A*1&z3vxyi;;qbB*Q^e#w%hB-gdM)Ln~=HipE<`_17sIJGOF^(^Ed0tYbNUeS3xO07Z^pQt(J*P5&>&MQWI{+ye=zI){ zoPAItq>z|(G4gtk*Rh86_2%CeD6#wW=_AiR^BmTfA282zLYO;GdiLn4Yix8>l=hR# zCMs~MRjUqV!)Ag07>cv~sP`<|B35QD6X*G;U!-s$`TMWGF=os3nMK9-GtWGuYc|Wi z|Ni^H2W!gQInw^Y2jS261!>gSJnP=-_m-}gb+D#XuYMz*H-qKnmtR85DO7<>{OPZp zyWI}#hwBq;p{QE9isZt4+60BlUbvn6{kLE7Y|D&wXskQ}+UPULurKCd3gvC?BFmFE zk3P@1o{Sm$lite-rE%IcX=Tf%EqG2WmoE_yV$h(0u%>xU`?(Ox*AvH2%H%1N5pcFU zo=I%ot!1PI6M>O1_ML)RUNUPPFHI?2>xr+dD+2y>?U?GO&4haYh!wtrZTpnu@QJAF z!#{!lWx}Tmi?6==Qg^Vw2F2i@!Nc*+_J_Rk@(XxQ-R=Dh!NIVJG(i+jop;REk4DtgB5V(NxKJgCgXXIq;If2=PX)+pJrPKLAKYR^MJXL> zV;%=5iy*s+KW}aVFFj{rSL6yDe|Fzt#-oRmxq)AYoGXQEC4{dmiy+Tm(hZ?*dTina zD*ymM07*naRMIG{EQ8j$g$zeE-XsjIi_-i?QTWKhxp^YF0_Mm(u3zLozGEsp2bA!U zr@4A$9b8j*5C?n;k(j9P&JSTZJH&qh_|fbcIi z__Rs@09B4Fe4Pvi|BHi9q_u0>o`=7$_x}j}lK{R^V#N>te0e#pcn82c9oL(STpw96(DX@B7%J_rPG#Se& z9WY$SkN>7YKEiZwt9$W_ey~rz^_Q-<{y-UiJDux>B05AU!!fde^XJX)PHh1kv>ZgQYTG>s8RX^11)E-gWhO^w(oSv}(Q=U&O~PrVux{-dJY+tEdz2n<(ei`*1z{9@qZdZG;w6eh zp!Bi^x$6LzQO8dn7Yfyz+|m?gm;F?bL{s=w*r}cs!eE*(=FOW|4ng303Ib*Z*2|bN zlN2plRD)VEwgrP@oq@1#CQLW8Lx=`>vHFD0lrf_eh84#t_^1!i3;}VmOPW>=t?^KA z-L^IGd0U0b_ukV=zQ@kN``X>FJN1`i=NMsehjxWZuvS4k^XARRE?j(sA!W&u6$*#~ zpj#<*f5qMXOR-Bl6XI;_h0yj*?1JdoqnG;pFI2FQR488_JJPG_Zej`ow{F?2yV)t= zzc$x(G8lr;-4JZj1fRSA@us10%uh17x~ZhNP`E(B#hF-M2Qo746>u zf$Ct1j*fyWteZjS-awtxFqwP{A$4->9HtwxJbCg+uADifR-Ib9+de1gHhtP3vK}7? zx5AYK`+FL+{pOo*wa)E#+<`iI(UoNK*=Jv9mnuW>cMwV>rj)5tAmG@IN~0K9r%;f8 zZLVAp3g0cYYSmT++Gj9-rw`D^w>0)*bdbZNqoZ^uf3Kdsr8;((&=hs}u;IFMx(XB* zd~v%5JFwe#cv!l@cQa!RJcXU*6kHC2z?}X@Y1LA?RB3gi#K3NaAp|X1qJ-{>WiWIG zMW-OPPTe}XJD)rAX&GVWkA3;l`3L4f*>q{sNi;sDPlM+^|KfAht1I21R4E9$3?Duc z)+1@La~^$zbqRT?VucED)AXmu|D?ys(OBM}yR}M<7RIn6akhc5pzh7p|m}&bj5T@qFuKH`F z6asg0-KbNiwmg6`T3{gL($01lf3gy6r zi9e%k9$3XhDZXe>ZbY!qm$9>U4wP;=pn%AkGba>nB_uiGm(aSQP~n2A)YyqRFdGWX z=1>$cp2Dxc{zgzp$sid1P2QS6{(G;5d|0l~DgIUPz;**{#j6 zPPA1EgcF#TUqWFN9UTp;lfyCz!fx_srOK7mYT`2}PdNW`Vs4p5G?~(y0`HpGokYPk zEk&r%JOKs4#9t;VF3BKC{2WS>1F#U|{Lh4S$F2W`3m2C9n9B^#$}zw5&X!t-itB>l zbABAJUAJ24bZ5(!Py~Y3@QLr>bMDiMVA3zYNEYzxoh|QF>x)6ay>7jmG>9QhyFbQX zC0K*d7kZHy$1Z44c1+`&?c+|K~JVhzy6|1u{4-B-$IE^ zrA1nd8I^=oTrvP+p@Ida2zag(c#d<1^mp|sM9PppBdoyA%Jo=BX!)=qW<9J*GT<3h zKng?2^BtZg+~rP14cEV=OP0x$X;amb5S{fB>6QT&DEZ;jw_MrsP#*rEy!qr4PpE>A z79J*@lkycSV2+=HmEw9`&#%HZXPwmrTStN$bEg zh&dA#6@~SoHhdtrf%VT?;9OGId_I5Zx`WS~s#U8hKeC_h`OkjNgd&fYQCUILf&~i6 zP`J4&1qD5oJGVA#s?T348OxL_2kV)$syI&%I#Hot2D`;L|7cZ6#WxdIGQ&6KN*P}7 z1PJ8dvozOVCcUfcpS{u>?ygtDmUL|aQs9(@G$j;#(@rP=UI+3L6F%t^NCB8v8vIvZ zeN|#$?HU~w1qJygxgS;{_doc!erW3lol+F9ynNljyi!+L-uRQBKndgpqxq7Z<#P`l!o4&BDR)AICSL;ZL4A1Mh0Jr1LY~RS`nBSln3QSA;nD?5*}) zdx;~4E z_>VzIrftCOC^sr`qfL`0aK%J{bS&_zL%>F#@f#rUo;GzVssidI<4?#2*Wme=c=TU{ z_u__)8mk*!Zg@#*_{4KM-i4W>punP{f+-ySSK(cbX(O!4FNXhUfz9s(230PBBl&~t zUwrtlfByj*M7c!qlJXp43TUFRLjT9j|2H?fS>IQ=0cj{U>X>rqG&$tatrW@tVWU%AU^-uqiWn*9eNnf(X zh36kZ$8Y`PMlmLCETl`Tp^y_WRtUd$DKl6MB{#sTvRr5FdGn98f^SCXLmxbUuVa#mLw3pkPC za~JJ`1@omeUJmI7h(VL}70q=B4~oKILB^Z2*gN;Bda!cuhSxi7B(Ne|Go-X(?TG(0~R?HkQFEp zjUdDfoQ<)7@So`{XiOJA6cke^PPbC*|LN1`EQuQ?!8rDafB z2w@vF0^PuyRD4iyNDB?R-#LE#gt`#g4r_xSNBs!bJk^x`$B&=VcNm9*@*D>UIGKOR z0q2am6D%+htJQ*nyodob_@WGVbTH^Q>CeS91;&>1loT-uuMWtk z3pVWz?j*PUNeRW`ZtPO81`8o7Ik+R;q!f|@v|{4U7t%EYc!}=kXq8pILM64RV0)aW zR5Fkr1VH*TVDnkvYw9OgYoC-TLzhbAu{;n)(;CXM2$W#hIRt_J?%jLj4k*`X3QPFJ z;TI?cs6dQ@;Fs-EnSs5c+9%Ezjz5S02x#D36;avL0ZIk#%#VqQLD^&~kf-}3!siR5 zH#VtKVLeHS{=y{-@^uXc{ByXb+oi*Y55nSXi=2dlg6mqN#*NSi)K&>0YymJ9Y>o5b z1$+SW<(lrI=&G(Hl%ce8VW9Tcv1|WVSb!|Ri@00=%a^MF1?27W)xbfr7r1l%I{~Fn z?%er6?+n0=aLz!`omP%u4XlB2y7szjq1YoUp4G8#4GW+#5Xxu7x45$%F*D0m!}cjSN4pVm>#KU$E`eIJwISJS?K z{LMWrVt8LOKyQ;}Z+Mi`6IWkv+A(tYyJ|u8U8Zf-)hGWkxuecK&4?R*QojFa2Z4-- z4^CYF_1-^#e?r2~9~A|K>3Ud&yrSp~96VeK6s`m-qkdR7AXvi!OJ3s|r#wph?fWND z=O%pK<3A?c_7dwc$KmteD0gp# zz96P-(p@q0FBE?}njAPCGBe6JH4LD6yyqW(W}5NW zrQVLe6J~Mz$thY&Pu%*;%%J&i`R{7Fp#lu<=)VwrZmg+^Kl)sM_W9@ds0o6Zm z-LqGm@cGWoHuw$j{BzH0PZ|7|DI9zT5a!*7AAMAX)_(li7skRc!tC()({hyjO1GD+ z!}p)0p#MF0wNwjv?y7#M{lmb+jCDf79{>wzzCV0DWY~G*zZnXchdXp2dF38jJzoGm zg9>vX$PXsPkN@bFas$>3HcxlA>M95AT!{ZzfkC91IQ|3RipRuZ&@B9fhyN2cB~70; z-4%ovgg4H^KOFq$pM754IwflpZMn|uD4&IvY6J}(_QAXXG(IN|unBadh- zr(Tfnuv6yy<9z%LzZzTv`19K2h2pQ^eejI}1Eux-Ct;p{26hBKX&s@zp+Qou|4H%w zapC!YCCop=o$mSNo_p?ir9J<2wbq5+tZ-HX%7F;vnkC8ndQW^?-jKswxNik_G`Xl} zo6;8WV_WiC@F1b0`DSDZFV~j0wH&3r+JTY|t~<61d%=B?V10>5HE6 zK)9=V%G7C4P*hfzJUe#n#)t0$JZ#QNgZlL~en`bi6(kx8fy&sa{Y>Yl^+oj{%$Ym1 z?|>ZuCa5-dW^aMgW%k_JD!_i^v5xZL2mRrmWDMda+^6K>Zhcx7&~JXq)F~0SWIJF_ zfgozX?vAHxmUZi6@KRq~CjI=g#?fG%;!q}4sagesPPiVb-#~Z$@87pal@Rj~u$eAy zSa%8h$T!3;A*Kh~KLl%neSqCorcIv?S4uCSer_3s9nEjN@rIUheN+bO%_-_+=U(C47SXm(2oM2_3PKkV(enu1-Cje>(}bXTnQ*D=+db3(@&{y`;~|V zP#!ccSfntlI;yGP&!%wQMS8ATvj%o89fBPl4CF%{O zKyK4N)FH=>g*%vx5UetGzzzt->3(M8hV|+a?f!Q6Y2aYSm*7}z;BGD?*5uBt{`g2* z9}@$MomTSDL+#=ErK|4XNCBZfh52-CMWqTC8M<)#7CXf|!sQcPQ%x8@UR@B;WmgI8 z%$_r6w#LDrQt7oBpH;|v0#*k6SfBXI zB=0#)_e=~#Pk}bA4f^!y4aLFJ!@A_# z@4l7y-+LF9U#(QI&K*&7t3{>O>b0x&#rS=ESiaf!Efp%$0*)WJJ$m$zyij;C*2KzH zD@{C}nDtQFOp^Ifba0N}2&Dzxdv(FfHphxq3tS^;B@zXzBDYeH{_%T_Z$V3(n_KIK09Y`8AzW460jn4aU7pjiT#xmb>o3PRZ~lC0QBeU3W%n@Ux?C7@ zm#(=e6sNl}1|&QUg%#Zz5stB1%T?$|iDzvhh(N*9Rk9X8{ENz<9s@z_(bg{bd zqMHp`J(jIdULJY)Q4J{n2n1*h?)V&Vq$L|IXqJJtoAA6UiFNH6Sd7pzY9`=b4JFM8 zSjHWNyF>CH6;dm}tAx$u_|MHhchs9i*cKok@lucGrvAG6dRs48B9Lp!+gg|T{t+j9 zblg4v{t@`V`hNxftFRKGqO%>I#S~Ax_Ih8b4Fv*~BI_aWW#R}sUeM9TkHep9kG|wH zPfO#M|K0er0{*lV^2u|_s7XPN$A2t?O1eu?&-lX{rbXU1XRL*UpK$QGUf%z};}Tt{ z0@k$&TK1J!U(xw>+pW!6+k@-D*Iaf^PYf(vI+!NcyKPEqS(dPsaAwNeS`L3#*;mrR zbw^VK^{y27UO(Aqei|VI|BWgf+(6tFf0J+UV!kPw+(?^LnaeD5!;C*K$sHsn@7XHM zapb{?ng8SuTBGw0H!IwE*WD0mb&y6k!veZzPg8?^SzqsNSi6FyyKF^$19iB`tktxFi;@Mn--Doi$FKG72Xp1WJB z@a<6qZ>JKVH{Ol-{X_m|iVc5mu24k|+%ery?M4KBr!cPrtOKg4TbH=t`|;;_eCf|; z8t0Re5B;LsOHDrdC&c*k-I4FV3}jDww1<%L`RAV3ccbfXxIw@F_?&(b7OL@q9|fyY zx~<}$cHTGs;qa%D%dIy$3cCJWnE!}>VEr?cI)HD&`0&5RlOaNuz;TVguNcZ>QyP_H zh8U&+v=NT~O#2Bx|AMJwuH3J|UpwgiSTrN+uO-AKum1@+|G2J@KU8iK$NIxTbmJFT ze_bx_;D0Fpx%rn*-4h1$9~BgAPkC73hOGZ=8j1HETz3q>cXx3I|3l$xCM#)0l(oe_ z#WR2b)PKewj}w39ae@YN=k*(Jy{Tnc+(E!;ASEFE+5Y7`{}N*Tjlee}eKG4V-#_(v z@ejZMM8^N0egBLP{k?A$yMnlN-yKahcu?0Vv*vK&(Y3+W;lbY49M`5l*Oa%l67l;3 z@Av-J3DDPbJ@w_`uK$WPiMZzZZzx*+pEz&Iz<~pGw=IRp^cSwltKyvmXXENp;h%(% zld(OxE0u1l0?7x5U1_*b)OP(~7{1+&Oi3lL5kS26X=7i!XJ@QE(pu@Q-6>E%8Z_ zGP!V<5lt+)tGzCEA+vS5OX6<)(-6i}*yS*$;(#AB-08Jw;X>UxUb}YPc#l8ZCB5mE ziHWk8qLhRmlMJI({vqM(>UZ}Ru~$+S!U z=MMdpSYMcqVeGj6xwXwdx@Rz7X>E`J?bFhP$R0)?=qil#=eVbUMH}a*wvL}Ou=*h1 z(wc$%Lj1^2ZvOfE#kGKVGOmtGa z4V`N*aI}1lzA9hi25G^z!ma;Y|F}k(b;j}p>KRI!mO1|6)}H`up80D9?l8#wIq*B_ z#4$LBxl92BX-CT{uK$c(!+FLvKRxKmG5HcM%38oekP1t}=ae}|+MqnGiMaMr85&5m zaKazp@%xP9$$Bnv&QK{jW%BP*yG~sd+yuHD+;{&K@#p%J3_L|nJ#y@buA6SHaP2$% z3I7D}N@t&s$^+nSx3<$Fk8P1J`1~XPr@{KlJmQ}c>jm2&4>MgZ{R91TXFlw_Q}09z z`L7ZRehd^`(b3UDcZJ+|5Ddr<(NWSC!L7^TKhui;*9Xfx^-hAjg{y?ZKhOgoEbr7i zxn%fLu({;^`yYS+G)DQKaTRLSzENIz>3MZg6POAHmi~HR`DtjMQYgr5{wUq?hz+7Y zVo4Y)lK@7hCZENqOhBm1_RNR6ZObcefa4d;UycIq%dv(SrV~D`ou6sD&ktM*zAjjEXLxuA!De!E1&rF#gU0lWQo) zov&DC{4vTThIw3j?U@t84U9kQhK|3hVAN!`pD0IV0$mKzohB3ARB|K9NLVJv2A`X5 zdSl}b73|y$Vd`XoHD@VxyGqxsalxlMRl1I1UHWflEJ4>N_9`6yL!f}^11kdE^n-pg z(UlAZ{T=bXOF!_In6bj=cLZ@}2~?%h*ee=?}Eqkqj>HPpv{YgKwFL=RRt_%~Fmrkep6 zq@J5l;>&-Huo-RIwCRcuH!g79(2fw0BREbSuFPgW14O@o4KOUy2L1f_vrlmjY>^Qg z|M~g<1#EPqRmas=XMtPN0ibU)h41hW2cNv&s8J*RHYE>D#0C~!=h!6xjW2wnU{d_% zA6;LQ|GpYHP~D{}(mEB1y`E@u{;U7B_0KSa#nnU9J4sahqobpR8-2J*)wt#}WOqEt zUR})i_y4fJZt7#ajERWpXqXSbK!@`+9RKkiUAs}?F=4_4ctv0kbn_H<@CTr6TFqCh zRzrS#wdUcrPL>P3V{Q7fM_PVdusrhkq!1J~O7njqd~~ zUM?1XZ1_9)?24ladle48pZ@;#3?kw2=X1wd5V5R-rS&Y^ySz~NRG??V_dVCai-CU- z8woSx*%lZ7#RvX{)?aNyf5PMM!LfxMu3kQ>y1Y>QBk&Xd{V%ck-y7#k82@3zhQbAU z2jREKmvFhxHG%z&J;iy`SH)M(JM~Vqoc}7JpT9=J?{$Z)Z*c-*<~$g~z!jDNeK`q~ z_nJBH#I@_SZrTo9Y7Kr!dRd7G-wCi&L6{eu&WBGyus_9xKLu?m(}$2$E4y~jd4eFt z(?eD~;jPO`XN8=8*gPkEBRq2Shp=B21EPY^)+&4s?lWtdcF#euZG>cWdG+UvS%^_` z7_lyDSuk*c0-+pXP!VG47>tuqxCvp^pNREv#{c#$hx-(M=!m>0ya^h#6Oon2WDA=<25J)iE?5h2SNqOgJLIY?FnhP1_X{o^KHR;=myp9p zpggnvW@~9Z;jtNiDnbZ4H3Yr>WcnD$G3~n6E$G0dEtpLC)1kuK6!iO(yT=wNs7;xQ zF&|9w^FL#aG(n&_BTROvq7DR_OpHlz{sM}ZP6)SMd!~Et%D~_idC|1!+GVD>PrQQ_ zAW}gyRyGNW?sKYKIr*1gtQ^CsWSwzv(U8P6xM2 zL7n&L1ROC}gK#6RI`0hq4UGfv*$!Q7Wk&oImq?!s8O&N_a52BUPj^>t{=4#w!@{^1 z4?g&SDncCo?mjIDTr%@37{0rA{?YL@TQrGQ2chr-e1iUgw^QIRGzG@mt-qFpfg7P! zaP9gTz@KMOm4ax2;uQM3M3;cBw?S(eln2eR6P)hF-1@6}b-T7wsS;3bxz^0QGAnyv z{N4Pc(&r}ximwQFkI@ARm`W&SdK-*b;3|0iVWkWqYo&wk@TUcgQ-rW>eg67LOeGET zbq<4X41Cjydln>DmwL>*eEK{54FvwWanjRx;qUb^K$t*t-kEFHD{a~iTx!XrdjBCX z1swwapT(!q`EB2Mvbh?G%X_v=Qec@a|I$!mqofU7LHzQ92HsS4`Zc zME6$uo{KkcX1g_m3e2GQUrc4E2(i_*4JU{osi{3OPd8lQzvDjxhgqiPetc*T9yDA= z{WJlA;Hd;L`L%1;R;ySZwE_F9cY}fAjz7hZ{|Lup3AAU>`cI_sQJUb{>)ROSk!1x7Vx zv|ymnd1tPiSK8DIT-qA0oN4dB2A}xyPv-;AFJu0-h8K=raOFK6-V3N0>DZ}Lyw~5d zaF06;o;m2ff(ngL{-Z@T`KK9T-Lbzx^Is7l5-xF8oAqBSX?EQ6&yPQAhr(wQ{`r3^ z6viL){}A7@Lt{}!1mMTE{%UtfXC#(n_xUH+*ytZ$__2+D1U?2MLVr$^ zi{?LaxULsGAq~SO_gwQzn|gsuTho=pCE7E$gBuQ9VO}Nh^VbtaxPAqIIsEwQ#8(1W z0v**;mNa)PCccBOtj#s;nmTORET9gWJpwtG!)oR{Dd0Q8JD8@^R~mc@N$FN4fQ`X} zzXlinnz5l3N=q#2#u~xpx^kvoDE<_7(}jx)@$_cGj6Z|)vR6z7rvLnZXZ*_}#ztAh zl%a4lnE%hGKLa=S#>0~eBr{^ZIi;s&EQ4o^phS79@LAK0#ih}oyht2e zGHoRteAoMM@VVps{s$k>o%D?LVv%(Rt#&E+bAhf0_gQosM!~InvHQ2-pV;H?5!4Pq zeDKeY9g98Tw#zAlE`|OEoNumN1bo6XFioX@Bm4}o4q==o)6s-ke`y8xEL@iPl})6; zzY%{OnJZxZbJR1OK>7TI6nQXR0aYFj1&K3qa`gZVQZx?)Dd*H$6Zdg%FFvjZ4;~?-e;Tg}5ON6+|1nyy>vJGJWP%F#nS-rVLBCWX><`{lh^FZYys1URg_gqj;c29s7$hcpSAb$JvZAZ^t(N+Ai)83R^XNhbFYib8wZL-@>;|#sfYT z3<(84zW6h?T^3jZ82%@+hTieT|7v^(P%&X@1@MnAeA3s9yP1C%f`1j@aYQae|Ac~1 zD+wxzly`jdKQ8$FAt+>=pkWAzO!uTsAuOZ|Gee^b^WUOH3nT`@KhlM+*DjX-xc*$_ z#`GfjpZJo#@g09Z|9jQ}sGm^uM`z-|pLf}VKr-h_G30YH+Piwnb2DcHHQbr6tj#ht zO&tRVWw=JzkdUJl_7K^+!jiey!oPV&#eD+qt2K(T<4m=tt>Yp1+D*XOm!gn zfmHfd_g`ryYk|NDc+y&2O4P^M4+ zL*rR8J`!QN=ijA_zaRfl_(O+&jhLR(2|+gvJooJL>WR^d~Rv!#d2e>^t2= z)~Hbn-8WQXo9_M3(Vy@aFI)^u#NN1v5-8FKiaHZ>3-y%#W-U3N{=x8dNaOMRj|>0c z@sGfd&|g<0<>`3yAEdxcx)D>){6CxvoCkqquJ~*25svbxE#hkUYngLhvCx_;;hHdQ zaUPzZd}ZxTQ_9p~-(rIgDWIw7HR-kEP8>L|gome@zOE~-)4WDmS6y%8g0q;}IC3%Y z8qtC*%*1jy=zVvR6!0Ak(}n*EeBEGg39cl!Ll77Ke%De7H~#+(`cnyV?i>^$9>6vmQ;r2_LMa2lb%k8i-o0eXO%ZuwB!;UdiXwQa#|GD*_$T_z6 zPLi1aZ5ND0h=k@7$#lvNwT>&#Nq@biKOBGbRSO~C8-EtZJ?3Nl&)Mo1iRoHN{1NVir4J{o{{N~I+W6t1+g`eO!SO~RVZ%&o4fWF&mL zB4nJwuZIjp%%w&UcGpk~yG4r@$r#36Lh9PBtGaJ9%Qt%r{QV#HI?ognpn`im)%(Ct zvw?A#`Nugv`0K$cY^Q}EH&D_tjPOe%7Rp`f?ho`3!YjVC%5F=6Sh zirytyhl2miaOp%vYUWJlJ%%_aeAki=02(9x*)RUt#1Y}hY7T#Jj|cbsdknE#zy5lN z?e9kPw~W%2@je;5r@R(8+Pi_$KT=L(ci*Xo&dHL`ITcRCs50HcD2ST8M z#JS`;#uHoBwB#;OL_3XZEp96Xs`?7DX?q#CpH%QI#3Y2zV7!bKLiZ$=I*R|6En75* zcB6)k5zhlY`hBz|iCl^={U>85WHii=^G8Qx(i=SKerX&Ynl~dTqxsm5AO5G|D(kO5 z|B|%`0NfZa3fHJ53VU>E<4=G1KO-k$VbRhl%CFHb%xRH$H(3!3=Q{}^H>O#J0%)XyyyDpmIAAK-7t|M8>08+)aAppoA3kGL%=&XC>aNDlhg;;D>^U4+ z{NT;IhN~XS`0w7m8$qS#$w|bAxH@wtDT!SxMyO<7tV~tcK*Hj$4>#Le&L75OU;z20 zctNXj{q<6~aFMY1vl;`OuUNiP>ej7&+4C>9^=~T7ZRw+%zN$@1T>MFoso1%5`qT-G zeSNbQE8wJ|KjABeVb}k}$3Fre%WH)G;QFLre=f}boQ#+xb~@VGfV}_R_aB3*McA!K zNr-<`R1}mAv=+G^am(h*=+QsPk3asXG0dh+nF&S7ETuokBnraBsHnVBu3Rac=Y=37 zuVkUS8wggolabx#eY#KJ4sP={bWV45uZJIe`ih>va4F)gEt2)v>AipdA=5T~nYaVH zT)Fac>#et{0)#tTT^j*jFlF6ueBM7;Hn!(apiN8ge}Mi5^1lhbqbz6!0vY-vRvsoV zt#|YZW)Cb$mlc|glHRq>SJbs=u?@<_R1bzmlcSQpkDy|}Z3PPGS0yUB^%>U@_$1`NntvdaowuZ!IgLDj-AV+K z1f_(3{-F5;UjMpx>n``UY9((VCeaXBPHe=c2Qz(==`)wEj_8VzuKhGw_#c1#30(C> zL&!f^zJ?`%S?JG683_8j!E!`n456XW^G8?t=lU;mBuDEWz{!mt0dke@JNqTF-$ zy&C^%?6|QCi0=qBYt__l^2a~_AA`#!y3gxBU;r#Ert2Te&^oI+Tz$O=9O=S1Im}wa zKYvicM}LO|abz9E%^ujcgI}ghSs6KUq{bJbB8GoVlq=P$Sqm#T6dI^^HkS3Lb?Y`z zFpY(B;D9PlR1t#mhE@cY@auo8*7q7-fFg!{d=tv0_&$GWUG)5O&ntek;Gs)wM<=BR zD5Yx$yRLW4^PdW$R^Y>7;6dK+3cl16r7JQb&wr(spZ_B7KZ(Qt;pac#`R4zH=xjo0;2+ynn|b;0p$sweP)&EyA=jk7 z{Az1ust}28h2OK4VA>88k$y-Ebq6Atq+Sj_uoI=g!?ys!VCeua=Jp1K>Pv3$T}? ze>%i(V4Q^@%43UA_Uzd;czP&&c0L^afBp4$SvY@z6etueyZ7#aG6aF4F91K>_)ncW z6#{^AvI0SqSFKttl`2D7;Cjg?uy)w=+5$146rKnbGc#w-kaAFLFkl&925y3jqKSBs z_y!N>r=RXDhYlUmfOubh`K3fhMQMQM+_`hBB7_gzuMqr)@e7_upul8!nmPDs5If|P zk3W(Yx3`cYMT@~?_5&$jzM`~h(?(LmFTeIpyB0V7>A(N)zyFrKdkRWLC_~)nIIKwP zi=n@wZt*wtG4T1KF>dTQso$W!FkT0Xq(j^R?u_XSI{X5K)AT=PsOze~{`yPmK^fu_ z;ZrF!e!@77N3eA1VwpR8u3UR4?0$4)5%Wr$0$(I0Pa2nJ2d9M@ArbpF!hHCy(q zlKZ;cxI&KJLOD~!tKo-FSf(ZOyMNza88vE@zEsBsKivFN5Ejmb=pPDy+t#g^FEdpc z;;>@Bk_x^b{|J0i-lPcqF*op#@QWQ1ga7}{{KwVdN`L>knDy`9Hvg%JD2sQIR}h#y z1_F}>^XF?2)5%jNLpib#yR$cHAZruI9KGk8IM?A=6S4E%)OR!AHSY2wobNvTYu4;m zyjzrqJDSYk0!tYd*He3KiL2S}p5DTRUgN^%yBeBtf%i{#KeGM=)*sS?!^QQ7ietVJ z>N~tqY4Ekcyno|ek>oZsW+nwVtjzx$0bZaCZ*fN|PH6M!cBOK*yJU7tGlya$1J}6X ziopM`)1QJy^HriMwTtGzj@V`13a)-?RIhHhlHXNG5&jFFe-Q1u>1gOfIw`3Pz07{}BI_@y5TL(q*ilzN3A6r1MMcTSh)Kk_ zL7&|y#8}T z>fs|tkk7>XCdIb?mn~fuLCcq7bHzScw0J3$p?T2XXnhK4)mWZCbc5Hbb!**>u^5~F z@?q1)Rhh9_BP-n1fyszv=<|pDHSpP^;O9>iEKZJ~zwudrs1#^&OA`&czH#Fwg-weU z)fVZqG4|<!|bsT_&40ZGzQHDJ4ikkALX%m+`1bTPlr8mo5!O7~T1ukAG~>|Bf9y zNt4D+<-no+>XMAL>;%>Wi5%|v{~7qN6O?aMygA-exVlC}p8uiz7aRQ}?_c0AE*-JV ze?CXz;=cra{{)|#6)eF2nxXtpu4dxBE9d#A%8G0tr z@DRx)ci(d_To_EjB6ki8_7Vuj?~ptYq)+;F5`w&~klSx-AxDlKL3<0KsMslU=gh^< z>yyG@ou^KmKc zp?$`dV36hU5bB?U(&FZZ4Y3=ci)77qjXd4Cv;6tzEZMYqi#+}0Q<4(p2w?M;&8h^r z8tpY~e2YvNHxBJ4S0B6#+?_LLF74x}A4W;GY+0lRH1?Y|ZIoTRcFOO+Pe$8WWyAUy z;Cv4hB$?%BC>M5OXMgV8*U9zRNuLn`ns@HnC7*rzi3T9A*Px!z>f-M?v!#++wut;Z z@n@-0xe|8X-ykD~4MW{Lc#tnNoo6^)-}m+-dhfjxy?3IFP7u9El;|b8=th^Q(Yw)m z5G6`5N_0X9B7)I7qZ`J|f4;xxdfv>dxz3cc&R%QpweHWIF`7a@QTYn|s!EHp$}PM8 zWuesjzx-(m54Cbv(_(ZEp|d6O^XFe8=25%qPpN{>keNm%K1xQ*C}|eAQIUM8DW_fu z{}ich^Q#4?%g0Ka-BsrPOs0=wiZ;W+S3BACUcE|qRJ<<)3ZxQ1r;n@iyQlGUGi#g30 z`0ckj3Bdqj?GB)c%l*pNs?Z3fZ$e|BpSpd&x(h9fF|5@zs6v%M?Yz#d%FscjmzemU zaJM-J>YXG$G_;$l&7c(?74}a%Ige)XAE-ugQ%zc9=$+OEkO_ZWP{w1Jw0O!v-nsVb7q-J!sI{LRV9Z@CZ8D*%EA23_RvTZ zednU;qm~$+gV*;{$&p~G#uOr^mjuk!Z|rpCRZyCH-h5j(t8PP%U__j=2&c(T{!hKB z1Cfd{e=Tgkqj=HB^A)$#uK|xl9Y0Xi#;TyhavOt(;1Y4yj5)*F@JDPD(!=L({SXU@ zPVoDSm#IZaTRd3!Bt@Q!S#EiYFP)4phB!%Z- zXVpDYQByxBJj}O?q(A7K4L(TtQSpbFeo>9Q&v`%-ru7Sd2G`F;Pew(b0W6%ELJSPQ z#D8vEle?CeaC0p6*+O45NA1ylsnN8DLeNSR5h;2~^S6qUGUlzHt5ua8|*-u~_JUG0hrGO~R+T+7VCXFcyVOn3Av3g3eN zzoqKa1sv^PWe_4Z97X3vb{-O!d|0Jau2GCdqEoy|gfUZ6a-eu1P@SF`be*a68=NB) zkvc>~kb>|WC=(~wD~tN?enZ;z77#u8$rJYsq`8)Ge@%K z@W)$Y=nsWDFE;5=y9CySTZPM&7a4d!eLa7w(f@|qo2u8 z8-}ad`5XtLS3Qk?ps8lh>sBA^+072tbiz3}C69jD{(HyTs02&=2< z`75N_{lLU#e%>*&y_?m^!yBrml79^I?1v!LV}O9H+ZO$2Z$jsE+^u-MhWP@0QRz(= zLj_>1;zYt71A9;hL1@%I^IDGMM54529#s+TCm94{Ux-qK6OA`dsDE|?-ttU=3#J1U zx`Z)90QhtX<5$S710XA6j4x*wK)_3#jSsbnMdT_2RV>4UFqv#$NFr9aRw!@b__f;s zw4zC*!iGBC!mjh=ceSDxAyvtv?nW74Cf2-}#n+j^6Z<@e_7mgeLWjy9Lz}Sj|Hv+r zf1%-?6>~b2N4Ep<5~wJ|=Aa(oxIGVQy*oT3wVUrF3O z>P&y$Klo>+gf4nR@{ipWA4UOK83&{g{t`p=h^J7>GY7>6dgZIMqW`QZkJqfv`e><+ zaPZOlz%`SI=%xLm2TO*BsoiF4KXOLieAho&sdu&H0Ik}Y^E^%^)Vkheg*aDYB49B z+--e4`MVQ868(JxK1?aZ%4WzXC8qPPZQTxCi(VI23~Y;?FA-1eLMmllf3OD;M*8;$ zAGir_8^H);IG^I@$KksT?GC!dO5;z4hD_i2UfD_^2zE3#3SvQQj#?5iGZ-AdLXme& zR;tCuOstRmk+w0)Tke~t?w99a5$)AE@=NAC!x{-XLH*LwR2fRs=3Y2n7Q4(&O25~5 zR5u1#!OR)=s_y@@0EpL2hB~6(6Eca@$bC>f8sgpV zfBakbteZMIp;#`MEc)iB=yP$S9wlkuGp%grl2G+8Z(X)}XeXN^p4Ou~lr4h~$`TxC zDR9uShw3rE^5C3XLN_F6KyO&BbPMCD2sI;vpI_Is02L@AiBUr1Z08?tw_P-|t>?=NrPppNy}9VfHXr5<`&J3C#<@Y>EJ?ya4aMHInx`C^V4 zvkH$C_8t0?_GTKyyt;-pIVF9M2kpktqTiav*5PSAjLfGvn5%*?GPw{6t*AQvARn{w z@VhlG;&mYSc0Z**oOJwK{(n%wRzas?-ur50v>Ym>k_~jQ#0XK=dLG!lbl?F;!S?QK z9Y4^TZ?^)0zC%EP6aLVDI2d?Le7w|m$VvH^dW0dAT&r`n><=98x-hm|hvUZ&+-(ka z@0a0w$%j{yl%iBxg+}^peer8U1eQ-{-_Kn>9}dOg^SxGox<}6<;wqKS;{gG0yWzM= zg`4Jjty{aEa(n+!bwQc54*PlkRchGPS|uCXh)uRO8EQK>OoBI>*z zW-cDiUuUI!#T5K2-RZ$;RXP6R6PwtUIdwNh^+P z?C@L$7&GfP>C%Y%rqeT1Ph{Oycys~c>5j9M+7>k7vQ(MR`(_j_u6_oDpdG@e+q{LO zsH=BXs+3|vc)1xQ{b}yKqIi=wB9VXkz4|#gv~w~}Wk1xi>C4(Vmu%APc>m0!zx|=X z@AsIogtCn2sz%p9pNXMGi4no}M2?9+b^(If1Y9JW0BwJOKt?12HT!=sd)K_$G8mo1 z-p*MZ?M(JXeX04a(Btjd$Q{*f$7bIt0ir87p8nm^HcXAo<}jq&qZW_uuZKwlWNN$*>En46%3{4^_4*01 zy-W`765jLiuJx(>c#CvX*0Gyu0c9I)`={G{H z2zx=>&9p{j`u+y^7BSHBke}hQqgyd)3r7=3|&(VT!JzozvY;hi`fV^=v3U z&sv@9YRVARvS!O%E&8Hrj$x(A?axTDjD*tL0_z zQS!;3p0|M!`vX6Z8(Q%DsQz-}pxs2FrUL4OnMPQzcLd#e9f(vEwIb_EYf*(eufkl- zxP>{Wa`+5!Q0rozC9Xal)006hdy=U%Kg4)Mz&Q%BYrh_Uel>7SJAe3O*>S8HoaewS z=lx78&%f1YTU`b-tguu!UWax~|IibiH=`E*>AApaWU580L)_dG#Y2&WRt|TQcn^^{ z3!JybjoY#@&-jb6iGbEmB^GxqCXCIX7-5H$<4v+;)QT0jdt~DKzR4XmvGA=$H;8gX zCtv1IupWMHeG9kCT2IHV8tzZ#UdMc(Hwd%glYl;xDmAx*=l0!NrSRnq&pdBnpA@=o?fcbn4Yy zk``IPryPC*vgezdM@*B@527&q$s&~8AlgRj@C+eFb2X$kkfe|G&$u~?*O+^!zuQ2i z8*NEq#Lmtao&W8;BARI6y4>!k>eq!6EZipL6%|$H7h5Ek%ZnEZ;Sc=no0XBTUv6X^ zHicT!_5M00G#C2kac;lw=AXq7!*()PhVB+jDvs=0MRUc*fP816&;zl?7}r@+UYJt7 z3v{P=uwN(-_Cq^{ekD1*g;O+*aY17hv`Z18Q3v~6y;lpJ~yAJ4dV+x_3(+jV>+-~No}j-^gwYEoA= zw`uODOD6Sm@o$HF!(|NAdFFxqxe^B3j~)x6K}0QNGFJ*Dcdw5Pg(*#c?S|}bU2Y|< zRY}0BeHqKy^0$FoKSacgcXU2ufUQx67#r=)({{l>7}=og*U0^QA(y8PF_t0ZhmwLi z*V=N6eiLOqbid{A?Kn?J`|CUTv;cwjb07ueP=J@8!Dh2|e^Qua zqf8(2G-d344&WcV^HOkam!{umrrUiKT3u?5bk@pWJ$n*+4p2toAboUST*9~IFllnx zjseR&FGxm8R>+yNuYXfO>H;|{qpN1vQ-6c^azmQ85kVVS@ktsqJ zjL;|nr^Bc%fLmZhw5gNQ*mK8a&$fP9wh%LFD`W@jDpsMR*3mxdl(2p3Vbt^G=96+r zEitE1jQq7to}Yj{9qvB@o-xd~W&r|rUIjwNs}49E=7Q-EqtaKM;T?;&o%3Fjjzojvz zp7(%Ri|oD0RRgV$Ss6|-`J#jW$oHer*gMQ|m)j&(p~cDoBKPj6M+B&^;n~TS`2#tt zOEktZMqnvkWEELkein}ybQR2^JR_qh*+y&UE{1scN}8zJ*UldW z9Sstdq{Cq*G12z}Xb_6`wg4O;@q?C^#Cqn(E=$4CRB4H#lStn{&~0Gzx0s7|jiK+c zSdnq&H(stycmybAmpg;$mR?M>M5xt=5u{!t19s@!>MBL ztg;&3gtkzBa$Zh@KP8vS8v}0*c&4h9C z!=%6XYIrG5aD%s%0>3-*xsqaW-|C@WS?LDE<+4Ukcx(dX?q!8GQEnV&omwtfAVMZ!uE|H3n&#- zhwglUbuzkrQ8mAw!7gCiz16GIh7MgCVf-OeH#_H-MhLw(l60tMjBHqsQ94KwkAj~? zL9IGk>L3bs@fE)MXE@GoPVvHRD|R|-Si~9cp~~f*v#N5= z)5~_l39r6xsdXsCQ?h$p+MtNj{x%uurTO`7&gGzLJ@N8+`-8d7n7?m__OY%CCA~6= z{~RpOIa$*r8n!d8vS_%-yckCg_C^+mKzS|uoG86pby6DRHeU}tIi2-h z5v^6c;E};)5&6y*~ ziTVNjAMVB65qD@$`J%ol%rc}gO@`YX*Ff;S=Vn3-jRh! z{_(x%5@mQQmPDeZ#6$81>|nG1^8E%%Bbo|G|DAHld?h_H4oISlF&sc#&(sBjqa{B4e;)F<2;I@Ydx&> zM1!d$kN21;+FCyOUQzAWm4h;U9fJUKZ{DHFX&l)R6UP6r=yEBu0V1G5#W7*bbuWylK_amq*a3QQFz2tN>@_RaB#fzrJ-T-&}4un z{hwZZ_=$f6m&7!`b-Q_w&BEI)KQ7zz*ocAmp0DCXN`1CRp^7VYSZLB5E-uv=CvP*w zLUn@@#pnC<_eud@-gFwqsPKE|VOz`o;OF2$EPCF14f)0w+zCTS~Fd_;o?YmRF~a*##p(0amS1-K@0U1KI6IlXx6OldhDGAUhca? zAGMx}Tva2j#keSqy?|VlyToz>6k6DG5FKa(9XcwBR0XpNY--?O;;{ zo#<}v<18bwSHr&y0ZyWI`RtCq9~`g-VFw%^>!vV^*x16R$BWPXs7f7g#rREul$RhO z{gMVw*Ts?DZyf)x&eXjobGoU1Nu%rg$6kDq zlUtx!C>e7G0pop=2mkE-=;wp_M%iU=wj4wa>s$D&yj#TQsM8_44X_|&A5(>XppZN{ z7L;gM)`y7&9ErPP(0StM<;dW4kIDPK3^=I_&G(4CaNr}5`7${B?@a>hoMop6p@oIn zaF)n7iQO0`C(BV7idMpTpPo7kW2&t&dxxls=9 zC!CWGjc}IL4d<3I*nmSDW8evd8s!(W!m@%TQce9ow0a%gOc5{d2R;o0KHCzr#bC7$ zuuuHjVBQA^#umzlemf&qJ9pA{&q0PEIu4c_Hg6;O*Y2ES;rN?cJ`SX6Fznqr38xks zFpwMQ)xB%O3oM4SoBo5_!)9lIT0}4K&RLv_;DJ{*OVh~TD!#LJM}N8^HyD7YFyaz! zE3$pYm#<(OXn9{hr(c*w21j{?kQcp*7d0PRKg#BEQW^+_-TI)4yPY2E<->I}X4X`2 zv6vV-BKA|}P8l}S&^pc2Z(GQ)Xo}H3WqfOyDD8OqehCmwYY&do5Rl^8+Ei={u&uQ1 z;$Aghpf0)q;`U1Q-*0KxVX8?>#Y+a+eoB5s+IWP8W7wtE*6wr8E-Ow<&a5rm*(Bf* zR$u0x5)C`|x>=&vhVAld7OUfCVBTm=ofcx;rnqHmSX9;7A-rA|b7&|wb3g1x2 zy$d+G$Gse9Cq7=s{+0(MG#PlI;%`uPw2&O$WLEMg%N9AP7!Z4gi zrqE7(02HTy1~DQHu-BbWZ%=Zi@-2UoFxIP<#OsJQ^MwrU5}x%IkBZ9lZ9xTf3_@h+Dy^g3n4`l)T6KmuKEiO}vm=E;Pwpi_KkOMs#<@`T`tri%yHx zV;W75OkGrGvPR~)W|SqfEQF3b>l9oN-`OzdB@p!wBQ0tGlRd2yjsY>JS(H<=FT*sD zOjs;zMR3^T(Mr^dMbG?|P`SnCc$`|W zD&v9o3@{7HwLc*eB2Tl!OVy%>x+Kn5Iv3=ZPxNt{X(%@z*di!v>18haH1e@POn8A$ zkGlRO3$4T(?g`i+JPn0%gV+*wD-LG-kz%HGSp(KxmaEs_Ut?a!mMZ6l@*I>AKaqWF zAqL;2NKFlm;QOn|*&i*Cl-tg4+GPIsP2sEER{@t??`E;bx&- z1{;n+0QHc++NazoU#{rci)H3{(whQ$bhcK)2uIfYvKOW0a$#OGp+WeO%bEv1&q}>B?lgwfl!5F(Cju zqOI8_Q*IdhyIl%Pj;X6=3pOaT^C|HOn^P%`D%9mEMA@}@BlJVfd+88`3!DY~27)VqPh~ubfF^yR#&)lU-#b##ncI;uR^E$(56bYzI1f7% zM5>H{SWsxst~S-unZjEAYf zF#VG2`5{v10(jdGWI&+z_`}aPhc16_Z0=u&612-KcjujahH~C{w#t zaUy!8*gqZW`LRi2vwZgcw4V5Nupk~3uf7K7>^cW!%`@I|?{^<~U87Tp?>hd%VGJUT z4zI~ndV$2;fbFOg+j6_M{UN#%FE@}}_#UqnH1ZF=&8oXn_>oY*IxrO(MQS@V1Bl<8 zkn%*BaLR075|`A=HDdU|H5^FjQ<1uYK>juJBUX;hmd+3fsj?l=*&X;-?TUJ#AY?i% z$(wV2hsY_e&BFTXR3-ErU^@hqFrM65>0vqP^CA@hjSGo)aP>H`>JgWdm2+6Bxt+w5I9<&C_(V}a<>3Z7fO%{Kp3M*sWZ)G4 zfLf9!yod8#Z}_kFd-HBHqv1+fECUxZ+R(us-K|4l`?NN68%O!58s-K5A;7SUNV18d z2$h&iDM!uq`^B)wx!%c;Y5cwL$KTfZ4X$r;`wF%mu1_Wb%q_Xn-?wclUF>3ixs$v7 zxY!&U6#m;{H(c7tCK_u2IOKwt;rlWmD5-PKP8P?GE-rus&2DO}(+{rzn^59jvJo_O zgvWKlh0`{CI0d=!3c)nIP-?u4Y|&T!E9xa(2rq}-mi65+3vYtycpWc2-rgp*a@MWF z+-Qx^)X|0fjPq=1-)8?z5cgKCOKan4Ry2>cq(?8bkW0S%_{t9ukTV|>;PzyyOWN;j zZihDH4P`_yJ~{Ft(-#T!Ucy3`_L=oGE#b1T{T*|9ZpRB?)b9QlpU$)y7c?JVk%$xWh|3vI>(S#^ zoo4emueSv|LG!4|&dU(8sD(HFT@Y%f%XMtWy`%yF+J7#GW|LS?(hskEEsbc)^S&y>7vbLX4y1w-8lQzK{k>oop%}^qwc~OD`8kW=xYFZWKPH&DfJo!3v?daf zA}zWq!_QK9RM2 zY;)t6jp+q5>^IY7{f=K@RKx4<=K!|2q9ewLdCups!18nEH-9Gp?@(c_i}((k^^Sf- zJoo{Dmb&utFs~Y7vF5)k13PX^txT6_wG&PZRKi~vP+7kUkP%c zG}8FQ)RxYEQ`2K9)>zbGmSt=g9Xg}@%GkwCax32T?85PWBNOnibB1phLZ0wn5^Rk} z4t$(Un(VjL9$sMr6(m^y_jN~l9hn%Xkh~`R&nPAn&(i?IaE&XGG;S_rBxVdg7qS>t zN`eZJ$bVPh#8dJj#LFw@WY$|K-Iv_CO2&e-VG!EpqE5-dH%beOv$v<>&P?>&LFl-| z@1W>WYOT*Ax*Uo6X)(mA5s}=9qV8};gi6qdYhs6@x3KL4lxzh+FN5%d2j=YR2`}>= zF(PI~5j;-}$WWVYpdW1POoLT3`|^ceeL7os-V3u-5!n=;Fi^1(ZK!h-&t{;%HBk89mY#&4(C*{etj7-=Cn*Id$ zi@a!^7zhsfX7ApY7P#QOj$OP@w=`XViZRYxZ zV0gmWFVa`w`v*c~8>tAki_HU}ck%dp(jObdM36Fwq{>b(Ju*%qTpSj746ZAlh`VY>r~hx4KqA;=@IOj7ZqqBc>S2rk@u!@{$y4#E}jd%=EA$g z5WX>JHZ2S13zvS7m4!QhJ{@e8>dip1pAmvX(sa-st!Rra`2E}-h^LTBluAzpcT!R3 zJD@_~Kls!+wiNO^4789Sb*zst;_|AA1B~3&o0YW&@#_iHCMudR1C^rc@eR|jwfy5i7j+t{@eHSa z+Nr4s3ZYaaf3fO(=6psbFkZa~o|eVB{x+!pT}~!Iq2O~%m?Ng=OP`aSX11vb8*MIK zMEvfjLLfKFnR-0q+=l2G~)}`rCR4 zMggFjc(g-kwD4aEG970fLuBEZd#LlNw+M7IUcJ!XLXOLX42T!j#sq*18SvcyztHB1 zT)UgK*?QCnrs66VJ0*1X5c~8XL)y6~MyKW>6Tupq|E-BX4_U7AzGVi-F%EFhaTu{L zPCBI6)09UHL;+^W<+x}v>w?$O$!(;v&e3op^c@7l7DTW?K{%sFCn?}zfC%Iu&D{H1 zgCFV*M=iR$Sg3H*p;_0zg=P+lv8&7|zptEmuUQT4`&>k5ea%PE`jZLSea=ke=wcOg zyknzTc6JYdXmb#03x<%(iHZBagr6h1ql+ML`xKhdFw$rTcKb3+&Rpk#G5j!%#rm$r zyjjbE#|aG>dJIn6wG?Q`_B)JAILXEC-^>5i!>G%rKe8=~o;@yPsud-DDuLSl$lk-4 z)u~QJ2@J24p$vDlVZi)zW}~GE0P&muR>b9SVAr`&BKqX>?7^rP@&$_b#~Hk(kD>QjT=gh-k)cSZ4c=x` z^(9kn+Ngg}#mwN)9^4W<&?642nU3>%_VONU^Ft#8Dqp3GF(74N+dD$t16u04c)HZ- zGF{8lkP|(gGAWdB)I2j-ieO}71DphxC>c@sO5^Lj0bmME=Rze6Z7i~n&EA)_BP>%B z%;m#l(X)m%yueRp@X6nJr8x`BL87|Xt|XdR=iCW4dhxshs%!EX`6XpBn03>TG*d5o z+swGIW-@F48}QKPtjN;&e^FHZzbIl~$Ry3C_2KZ_D(8|y#)?#!v-X5pt+YGQ-R|ah zK9rB}oQd5d2!TlH|B@>!3m5Wd6WniunZi^TQB;6(e3XB&+!vf776rgBDu?gOmQ4)( zYo+>9QZo$#F4DYPQqI#obncBP!#>3gm^rQt-V3DZ3|qC4we{@9^R^qD!w5RD_y29| zbTt0_9@U)*<51L&UH=nY(KJkx?8bd}jS?#X++nku8=IiUVw1~_U!V#wSnLTEOBzU^ z^8>PN9QgEgXy#WyjKBN6F!=Mm{VQ-(w()5*XOsONaYsc@dXK`^cTJiSb!D5KLTPX9 zRh%hzBD}-L`Wst8y#I1~DuN)FJCnv*)*ZfoP0U-v-IV-Uv^B{&bDqok?~eXiaS3AV zH)IEtiS$IE>&c)fEhAwn#g4G^ez5K*gfxC_@6z$%1Wb_{35qD|K8^7MTOWb@Lv9}rnU_7Cy?+&r|{f*cL&(^Gmrc=MkzWY6BmK7}R zdygGV<16a_ycb1_{-N|+8X>l`X%M%b$G`$%rxM4zJKFz3q^8#Ia-)KIDwGNIUoi%td9kny z-Ue^X0E-y(Y!Gc&(OBTetAuPchM;f{p}o;r1uHmnu06QouQ_qkl20WkUL=jH7xXkuI;H^LE%(tqug8d_td{jT(A z1P;|2yYGA#0LBb-zJBn1*0GBX+}$*3_Nw697u;hbk9QZdIS66d1^xqBrmx^e`d{uyHE2k^d!rmoNMTBA`Ght~*o1`7LV2(^ z=mV^Je~U=Jv{sK?+D-#(EHZ!cgo~+GQ~)%9S~RJsQ-TS7Wd^{)gw3PH#eOjm?mDZd znp%^I)J4d`hzW|~U};%SwKL=Rm(~=LPjg`*LHCV^Jg@o)cs<}7LaYi-n*3MzC`V?S z_l!$h<>v)81&$ssm@u$G2R^86Tj0`tS`oEy*H}V4)(zIIXwXUS6nB~UNnhi+p=9h% zOrH-JNO=1k{t{NxzP-mRh3q*ezI(#F81LJE4$HgQ0I9zN-8YKrB+&#n`E7vw5n^_& zfBt;L+Oj5OVLD2Iy_)*-Zm`rvk6kb9;A!PDOhz-*Yx`Z28>Ui}P7FA7S#YJxC#KN` zZ2ir1h3)b7Ta)e4n2Stc zB}1~dPxsytbPIeV_Bv!0q69@1_2-&e6{DsP4Do*YwFJf&O4NjFLCU3 zp`>eJ;MqUiBG2rbxIzCz-wD^c0q*gyVN;KDJ+yv- zkm>B~5TH^A*BG*C?YqN8xmr(|Y?J~9N!-I80`^H}#4#{@YjH1BE0W^y?r<8O`EM91@9!lB4i%rasb-Ma(s$kK5 z)&<#l{zq`3=C#<&m2JkIk)dI)?D^JB@x6ScIFP!^OoF?p97kb-*$NU-*x_*4CY^A@{Y(YrFO=0{v_NZshmm`$RC>wZx!${ok zD)*D$J06m7;q~apx%n8dGP&}}|9Y&^Xb_xp^?)!t1fLdKpHAE`=RvvIe;{*I!wSCM z%gr~dkwSa|lqp&+46KIpC@O?At1bobawUe zZ|_%?yKp&>+)toNBuc86MXc1h0o>4Y#7^;~WwN`cshI`eE2D6#6$Yau$d%3LgA+)iw*rMan-$ch5Xi$gUfXT1_eJEKM#W>c2WcA`(KIqq;uJ zo(N=;YIGrkNX~L{WJeG+LqVNH>I;x@8iioQ(i^YrL$T_@a^N!s%GJB|@Me^c-+aUN z4><4uq7;0#vhTp}7#BD<=RGTbH@+u8|DWxpsOzL^_@&09*M-+{jn4P&TUm-S4@Saa zq5k{wFtmq6Y2~L4JoJC>J2yD-sRUzxIdtpfeUvO_Lo5v6X2Y*mMd#U@F|yKJ}{z6bu#bKdh{?aiK@NF>wPAOMsO zvtbgon&8kdN&8ua0moLdYSEcvXTjxS^38K7J-`aYwHNo>umm zc8@s{wWsx2Qhaf+DXBSk7srGxc7zbiqSgYDvsvph|EnnF`o9{z=DcF2ncY?$2HEN( z*S8%)zrvgE4)uVG=V>VXqzl5Cyq#7T@n41q2>}V*om&4Zs}sGrfGn|kYKUsk-N}=U zKO%`E6hXW1SKzzZ_GO{pU4iF>u>m)d=Js}dPqZGA?i9bi zl~fS;5;~PDBy@q>U1gs@)YE#4vWI_$*DuQ<_@;<|H1Q^T%1kS}-lWoDBE$S`AEDM0 zZAMYn3&*MKY?Se6{W~eg5nq`a1C`CAlS89w?QaY|^DRo~#l61!KU3}+rsx{`ceS?s z3XA!=eRs1~{PrfPg365>ED8CnI*;8p!yj+8Qs_ER#$IYY1>cn_unfOwnKW{)zZi}u z%6`Dla3Yb!(R~zm37KUfaRv z!7!9z>jQnLHfj{oVd=yIlIG&75AIhT`*EOY+-PiYu~hwzl6JKZh2v%=9!#K70MjTiQj-+o{CCokOo_Bsop@lb;NPO zna2fb9YpD|+M_mUOE8(mgFdoOXA!&jb9a(c+x4GbzOdTmiyczlx5takT*1EkXt=Zm z3cfonI>n-td^KsMvx{Gq!%KEX{m3lB!lIWt3Gn`Q2;-TIUAalr^SCy>{cdX7Y$=c; zJlZDgxr>(M@G8!;L@qd1+_=+|k>)kgCxKk1#ngh9Iyr*Y zFJPDZ*k68s#v71H5stwoCZoeW>p%8fiKl#gHGsLQJ{wuk;iippN%K$U)Gc^XQL5T| zv$R_I;nY&pZ&YU&oJh$x(P$x7JKH|%k6Z%IW&I$Gk*`Za9eET;Ndf@RP29l*zM~$LC;KFB8~DxoynnL5ev0qI7-eiDYXe< zGsa>{d@i+VkBoW}rQW8@mB-GST@{H)c_GPZ%K5qg=kILU&CdcE?RRY>nkXg8*LM+V zJJZ?NOxmH9lHNH2p_2J>KCMKQV;6ma8%7m6%i4lDR*>&-OgwHqpMR2rVv&i*qE(Z% z9xSg_2ivEI=b6uI$KUqLVTN^LZJWI%8h~`W-DvI9;9gjCYf;|hmG0~IKZhl|{I*nm z1o+)DtaOKe|2_2Q3)nlIN- zl7#{0Rbd4QMfQvO-_@Op0^?a(yL^fjB9N#KL$F==M7})N5vXqRQS!^1sUAXlrA+Ng z&0pn^+!qu1vecsvoDX01j$-=VO`Cam_;OX0Fw!u`M<{EXEpLttE4}Tiz+|5*spnPbC-=4r$F@~c~0m8 z$uYmO1B|RkN}oa@-Ox$C=Gw89aAY_0y>*>6w+XAB_Va!s-k&G?oCnK^%$HnY#Od6m%`f zfMf6us+d?dSS6@Hne42w2LAK;?mxhg8Sld+;Z`|6xHXbzlcnbtsdqW)G5*_Q#kX`O zw8~&ueI7Z65cPqpcAq^QM{b}G3|hy{3X_U6`RUf1e`VV-4j5mI;cb?-mUHo7)uZGy zC4j$C^3k>+Aud`4U88!4bw!0Q-{b9!+43X=zu@F7+)9LZGRpZK46yDZ%^~f&r!Oy8 zBb87IA5I=rz>bbZWGK@^&g9>wgf8Q^oAGdSbZkCTjNklFr}f(v!sY^6;YP0L zyA#GYs`U2vA19b(a`w34O`12J@Ii@l?3fga*iUq-_he|l zShoA7l48!G=LcyFhoC^q>mjgOxJ`QA{%NhDS5YdrZbDQY&bMF7v~-a)5Am0 zyQ#qeSFrFvR3`v+CMjxzs0#|%U_N_1rT_!Xnk`-Tv>(5^ap_nMKL|O2bdX5rd}L|P zHBljykn(mrhWN9j&9j0ZQaGvCZwZ|IW_!0xJSP9r8{tv945<{xo$m7FbVP#E8=4l|Jb&>5Ybe;D)-Z1pi1K8E_E}o_FAWV`P%Af=Oz&vNx4$ zqn^S(lQXuLy?a~cFxg6W&7jqil2LYq9SNL{&{bsXKP!<|itDM6YYVOZCG1}H) z-^=}2n6>}~u8_obQHKK(69fvEz>GJO%fZ8A0Sdo+3KaF-;dWhV#q8Bn$lbRIvbF}^ zNCf5kr)w|=8zX$C1TKFs=qb6HYN&YoP$IL*UR=jr zjB5@h>+oHZ?*0D9%2}>(6vo3VQdx#-a(pH7tD4z>k)58J_2cd7FEVWi8@}D2uKhD0 zA6!Csh~J{sn67vMhe;`IBA~A+ldB|%fUm)3;cMAemU?D1CV7BF5R&;}`>u6UaQ-a; z#gmNhUREIJQRwJ9#%Oj3&ADB@w*xYmWSL)M{~aFQ5D+h$nm2LVV)89qwBSf?K(D20{2#&BWpUVwqg5ID#QT( z^WI0viNE9GY1M>hs)SAFL$!vzK|a}NuL~s8m#xE1EIT|1(YJMAbfWuxcN>|!E~_~+ z(D(~r3S;&s;jQU-kyyK37nTQ9OzeK({(G41j}tZq^~>aXo$f2Do83@)dDvm2hNHsHYh0 zzJl0}hP&noc(9q^U~{BE3tjcP-aP#gpl#(%9yj}#SrDM=S3N8H|RAn!sKea8fZ`@m4fh*^&; zfR`UmFk*}ocQBY2T9k)z2xb2R9b8_-QCr6bmohuJ`Rz9EBk$z)a3-ITG5jBMZxz+n z*GKIl#l2YYQcCgSZpGcbrMR|up*TT`v7=X`8<4kDf|*8I zhvO`*&-0!)e_mCL@5(th^E*^1*K28Z3fD=H+|wpTEO|@v|3)Wj3V$Y>RX;D{c34X5 z?|Lh1>zv%vhcr0;0hc3av#JJqkSmwg1G*Me7Q)w+o9LfUJov&yc^%s&;ZYTax=Z(#f3e39iivBEW& zk5qZNNXOLVXped_kejjyes~p>~B0hjeJKg(tcFTL?RC}9OhJ7VP zRY^#Gvs$rgdJ#(6sOjQL*RVmwo8o>?KpN~*yJvdMr}hKYugmO07cW^)s9J!0oiEBP z#5}H}YVd14>rzfabc)5&R1kf*d;VwLN)_{ExRb#jjDO{R%#bZ#xZhGM!Xbmsn=R{+ zD~o>JOR48caN7kQ@)Z*Ra#&uShvK~p?|aQ)h;Ni4i)@ehVo}N73YaNj=x5{g{6olE zcF@&(JKK6ClJ+m6mOCbW+qey!R)t$@R1gdVjA_Pt*R*pR*VWbJ(F-{6keP!mZ3BXQ z(PQfH-gTmkU7E%qoQxGd5pnFQ^+JNNl(H&ej&=FG(Up_jZLN*?G`uUogor-?Eh`)# zp#8|$QPKy>bCYv%=hnwxBk#K427Wp=Zy8g_6|YQo;5JkPfBhqRiR{cjINEuwfXe~b z%iuDS!%BD2uln~z+x)FYPSN(T!L?|}M`VBWT32KamMh&B0*`eY?=3~znbD`{x!{fh zC#|_}yP$qBqiuuroS?gAF~2~z_C!|^?gY~MH}{xtVTF45ubPHzave<*uj<7)_KFmH zW-pHpA@0sk=o>zP5f|!*+HiMO&3BP*rbOe7lW-%|Om+e}{g2`6zGlsqIe(NE>cROi zjj+%w3V^@E1eKU)8D~U_;hl+&C8Ke>m6W|V)C+QQqHN=RgqPRk*(iJR)oh)F| zm2&V4;xW2Arb_O^NXBi3Ch#?6Wuc>G#cMwIz~3EkeT&vY`nPfalA|_FA7CNUeH%5N z@#(`a$Lc)U%unC4w!E;(jwEZl6csH@)vO}|#R0S3?`c{i<#AZy=dD+SQ?^0j3XfE4 zs^yoGWbIl~@No7dBe?<%x&HqMCV6TrMwSCn$w$B6yN31mF8|5xuvmDMNuM!w)(WB_ zP+;!@J2Z_F5v9C$0l@bw0u%PVy8K4brVt(HogCXNahYb4wdobwmn?hS`Kt5uk~JZB zh6Y&@UNJR(<}E?-68Ag&(C^<__64Q2mJBP#i-`-09-!6dobe&NhpPdQ4^Hfbb@%{y4 z#ogTW9$FUmQx)KqMbYV%ySXDa{n@`$KA*0WBzyQHnFQhIQ0|n!!5wI8s05pA7q!6E zx4GixMNK?53)%*~jgshCgfBixu;p^MYZIjHyz>hn2R(773Qg$KX;BfT?EKr`Qm(xla8maLKp&q#b749 za6*L1IFJHy?E){G90rMw78Nk9$fj!C2aoFQMg*!xTdfs#T8J3H;g)99wGUB5A`Dx3 zRAr%-zY^#NZ6>@;sx^g5=vNN=#HDZv|0h4H^5kQf(LLwBLkPM~qxtNR>Vb7M(yRZc~pZ{3On9!M5Fe|$63Kd>KmLuyZk#E#NMVAipbMKKpxJaLoFF!aS1 z5ciL{t>eLi%dr)NjJEkZ@r|a>`c-+^+u}lF1LJ0({3~yAV=7Aah0-;Q+de=3_Bk#} zbkbJkQ`gK}!Qr!B!arQx{$PS;jYZx*MRnMBz2rrcukKpTu^`%`ux$%cX8X_-#lG#} zvNGr+-g=8a1l~Wo<0u&Uzys_KKyrtuYoh6yX}LjJKZP+M^U;#&cXuTgwD@w{Dc2Pyl1MYwMrN&XGvzB*QCcUS_ z7nQf$GIvss+anKEA~1UT^8GFATG^A0`8^~>^)rb6bK;ahJhhRcDdZw}4O>5Wct2rY+sp?C>j5P7 z6;>XV8zo8R_X(H_66DaYOEUGp`DM(tCex*X`$_=c=-2E5ugwVY&0k9fUt;c+#*oc~ zrT)k8p@L}Ax8;n%pqB}fa$o9KHX*%CYum1_{UsjXXohm>)fEbM)K_D*ZFqhr8(gZ=$@H=!8z1m-O$UAD%{Yg8iuB>S-;064 zU3pkBEgz*BAx5teJ~sv4Q?u*5=E3t7lqqpEj7aL0G~%f-XgygNFvEf+fd|M_lIxY} z%Zc(n+|{XjMOL;UTr(ye7etND^8Rx764@{^vJCxszn6Xf zhMY0irzqsxKS(SS#f?s$wkD7MbnrT+%KUiy7}1x9uv$GYe9i#BWv7K!6FEncf16=E zormuF@tHfL%XJK+ZmjvEEdo)=CCaf+*G$Z(3iS7U_6erYPMIkQB19F49xDu!AmVPI zjSq_hvAtzAM_Lg0_>G6B8jT{qEaZu~E4B9J<(mx3LcrJ$|!>sbklYAPFnkWg%O9 zsA~%Nmh-eARzG}&0&v=NvqsS@<>n!j(CrUYbAWW$ns@5n#V2{;9`E;ic#seE03?cs zA!&10yIaBrmB1eVt{Ln3UYX#Bz?<3SC=pu(6-CVRWLTY}z*^9YM1Y?L42xY_cI7Aj z15>*%u~a;uc>E2Hq~5!Zs_e_ZWFx!%8>h%e3e&35n~V+R(blc`&iKZu z=t622%v+PJStyvcUXjjWIki^f`De=PnvXH)1U77uXNj%oSFNphzSwAVFe=0JTF!_B zand6?!kC5?Ip;9s-chwo{#pBMtxH@sCH-=&KGBiKa{$u&`2CD-<}OuBpJMu}E#8h2 z4nQbI2E^M*aToK`Jq$$f2T1$HuZT^9d1v^_-Gg~!ia$Qy(!a!_09nRsS@_ikdni56 zJgsMkV2lkkZ?!f?&S6ovUfX#j_^0E$S;R2-+@01(a_*eyvD+}lo1oz-+>S<>-WcIw zw*mfr&ah0?Bdu)npPmO96%&$LyvGr@k($GZ=9dE!A}|pGpzF>Od(acZ#?wL?9UF)D z$sD&M9*inEI%*gm502u@>C5GE++3bI{&A-Nmr+j{*M`0P z+t>7u!a=Va+%UbHiz2?%2I+k4os3eKm$-F(lDFB!KXX#Sn>Ty@*&GZe$%^thS%*cB z^i!=Yrcl&$Jt#P%N;C11TrBilkS0g)w>C0!A%7fVm(YR~ZC-EH1GiaTvfrFTz3^Y*wF>gJLi6PGy_rU0Q20lh%Eo_)oU02 zUtb3?q$N-BXX8&sliTSW_vior|Kb1N4}5%{jDVl(e5A4?7w>slgpfdkXXOH|=^-ST z@*VPZTYc|^eqU1C`%`~T3k2h0)8BqRgXU!x2yWO|d@q%l{2{7GUiquQ_$W?=mw!`_ zQHro~u-svf9eFftafo8^qd8ejabzvtzY*yK&HhM^;lT;kb;HoT$2i$eO zH8drPNb#eyx^G<7FVQs-C9bqYeo_5FJWje;y#=~Koc-noAp72>n`2LmK(#7am4Pv1 z_RuncQxs;A?M?oG-%DP{o1fgvo0++Tihd6qU1Ym->Ogo+0{EO>rrTtb&;xKmai~eM_t9Rpn%k7N(JN3x-)3XI1cE_Y$IF)Z;wL zTN4-^YxS`|%xRBJlZLtn(_uM4paPa?AQYi$eV*IPX(62e1?G?8_}u}zXaR;KM{GX@#8rZt<^ ztt@zVUHKsq^(RymPNHP41O-LKy6Kh0vm4W-SVSZsM5vvX1t&7L=Zi8Y;Bc)V|R#&XU*lTI`e)%jSrjLzwzw1J~oBGWYPZw~}b`%9P(UQ`{HVe&Yf5OL*_p|SC z>=b$7N_D0o<7a{W<}6xP>U%uzKE3FmW!lE)2i`)ZLo>k1c&od0x>+k0meDZ(iUz&} zHOvKd()g>WL(KIe38rKM>UzO2>hXVM-+Z3c7h(y0QqpwwEx$i!@!7`sy_gD~ZE-cH z`*=x@bPzX6VM7M}0QfbXN2)h(VK*)21}3IROfUh`_D_<>H#*u6KM&z`#E7HGG!GJ$ zXsWQjd3Ga6o+Cu2d$anU6@M^JklcAIo3yKw|Fs>@n4Nmh=98%SGJ(mFlce5->R*gv zIK>(+zaNbTigcyUE)PC83|Sa4TxW~Vg^c&g>QA2P!p8Rh|VVN(E9-F=!#p_*pl-K zmPzpPCI<9#JnGC#i=Ki}Xs16kJlljb}`4bB+1;_QV(9GB}$ zb6+-Wb>sO&J9A5C*JR+Bo2&w<(K5_M^fA?Mh??IBl2@t+fHWGY&mVH z<9kGrjWMooK|X~#UwI_#4hFwbIfQsPWA5Qoa{+E+uIVwhapMV){OSmj&lqc8LhBsn zywp{I-u-%UkHJZa4mL&mZ)|l0Pw-~R9yN6`rbRx~1Yk`4#Sb4`VimqQi&wy4ekG16 zCseJZ3^9Z@WZ{PGGYaN;1ho^dxd@bKgkKcm$nLw>QUN?FG_y0uA-fi4^jP#f5 zikb2iFk4UPvg5mY+@D-PR1Ninxr#7X9;_q%%K83|aB@&B_Pjy%L3cU+*7GUaYU{!R0N}IsDJ)7~jBDqzi2Oy$#=eF@uFv|LMtt z7F^31gUz~#g)ldrq#jxmeaC@=S`I<5Gxfff-!@fO$RU@`rtLlxh!xa^llapa^% zzN1J~#7{N>;39h$htVSa9QlQSVv%%N=PtWNjiw#5sh5U=Ks*2+!tZwt@X6fyr>$cN zlk_v?N@V1mWn^UnhFIN|09V4hGU7%GG?|j4;(&1I> zv7h(n2D;r9?)3pFC9edGTR7f9KKqhtvb{7eGcUlpzTUt*FOsBRAz+VQ##QKiPiws6(^^h zei@}TahOm8hJI+pJ{G>g&a~~gZoWLj>#mLTmr5)7maBEt2&4u-2$x(SUe)`-o)}uA z+;N+qe2Of2OJ^Z*iKbCmMs~Q%ZWArZE-6Z@q+T0}MZx*GH65d$BJwu!CdtE;^{BL% zuq07V+yR&hmkb2hJIB{pzUhj80e~he%bdR#C}YOR5rpzC-a{tHg0rD!`_ovYB%xbvY?;#B~FubXHcj&3W*zA;)15j8%=MTPlt)JJHUj+AZqqkiN->b zh(-oc3;igG30UOSK{6yL74Dlb23bb>{=_17(5!F>xh z+n5Nm4m^o;j5Vjiora*_v6DKQVJ9iV1PG^jrm7tI1)&~f{ZRz#wts=MKB2s8{P^x= zo7yOm+bS#58%YvCC;G-{D{cNKEuqhr2zfI`I^L#_Z+^S4F`74Y zRgJrrV!ZGxw2Z?PpJCXoI310Ko`?Mlujb^tTjR`FYMO+O##SDykir^uN*a2~Bp9#=XX7lV*2SN~vvX$4qWw8Otw z$vdRRpCOT+xuTz;gmTQ9);51&Mvt)>@1`;NB|t7J1G4a9v+Bh7VZlJewD=Ehxj=J_ zbTq;!zrMjMmSNd+MYqVu8J^~G=0GB>e*zy?YNLZ$=%W1EI<}2 z<;YhhMN}68wdBb=T}iK4L%U8HT_El^T*ySEXh+gxqp{dJJlr;MN9b%`pLyhuvFVRgn;6~t}+6YzKydmug_O;l6S zLXYwzgtW^G*`l3gz+u!kbg{qi+OL=~VOQPpN_qXt>@7jKubCR_k2-|Xh>z-IC9S(5 z2#JWt0)5b+L)dIC!)FhQT>YV8!aAhz*^OQ~202WhgEaeGLj_q6lP_r}5jZ*>~yu92w{8qx9nDsG2ZF0A0PxW^9;>(>J_ zd-*z9^hARsQpuqASH>T$V~OLuZB~Gf0u%Exh; z1?$n(F~+y#-Pe+pe0v;zD%Dt;Qu8qFfX#0Wg%sV@sYZC~6DAb3PUe2-#ycOU+3nFq z%e}U3_E|Z?2-MM)XQjF#j;$XCV2Pa0%ZKwW4-_kYgU+HWc}0jURO!kdV9+GYUJ*g} zR8=e;35*1qzp)E>`ug9ZH_g!Y>Pxru|NVAM{+F0-+`4l$oRsdikK%qIN}@z-I*H31 zUrIf#6l5t)ase_ZEE>OdITgQUpDEQc_W@m}YjHNJiKmj42QK!r1G_T;Y%-OTs(WJ$krod7O}Ox@H%5Hf`bF_?X}QEyasPR<(@ zmC<%@;;T@|=6WE|=Il<`@f1e;zi??`&&1R7G3onp`oll!_B`d;}X7%mrP<;hTlE?i4l7qZ*WnRu`0=5P__!O`AHr2ID$`a_UjIgaYc!Lf5G5 z>g&v9*R6d>m~|zD+pki$!P(iR#fyF?a~o@G#gSSQ-|d5N7)(lMfjuvmwdgPy=%AW) zG%5@XQiZE-v=i=^@J7?2Ys#iQ&j+o>^u-O@Rz8QDWxI1N%=KHLvfe#f21KbC9QiRZ z39LBwXmue3RFg8E*Y159ajqHC=|SFPM0BpTZhjAicbG<9m5c-3HLjaO9W5{x^Q?xF zhPVxm#O*FmW+%*w>vgoWszPZ9x(SxK&Xf3~mRP~lFZ69BJ&Gk;%LYsc=W>}zls*;& z@S|`A;GODk&Z5{`!Zhq-!)Rue&evwZTJmNxc+x(Th87=pGNI;@q}NJ|1wcDuxl^e? zm_QScewN__%W2PPUE-%@XW%P7N(5#%M$9>C_BE+JdbgG?vJRw&1&8xB?K9Qlh8i~5 zxtG>$3XjCl)Oea;nluVx+ zgjuzI7WXS0Uz_0`5s}-ht~o<3zs)= zDt=qB%i7K@+;FuqE{=Mg%cxBzPwOvR88EPzZHj)inVg4pEm8k<{wT78pq~0$Tg}uy zsBI-J$G6=7PI>-a=4)igj7aUafzn?;ZrE;mwq6BlVB|F)pTgv4f5=lLp&yt5oO%HkB;|oJ;+Ds$lvE4tmm&Q~MN55EdWO-*hVv)gng}oEej` zJ~w);Xy;`o%)Zni&X#Q%0v)u;J-$SCuc=9$AZWvv%OkqPT1jroZ-Fmw#i&B!i;%<} z=>tl?L$ALHL`t>mMA9R9tNuGjmHrmicSbv41gsQ;W3s*T;d5O?B`*M?yT5p z)D?I&^z-FQ1ZTYNTFzP^3;H`U?Q!+HOHe+dP*V?P_B!laN#n;2j zl9DOBtY`$YTIR^4Gy<)*{Yx=Q|L%qMiY;0INbf%tWk>d)7eqj#y13H-_u;ja$-{l> zlLL)WIMu9m`FlGz&VTOdZd(8Z%nXc)rjtRg|Od#(s>C4uLmk{&Q6ye!Jh* zZP#Zb;s!PNtYx=TdVFcWW%mCDLX{`#)cmd+{?!W3AYl`w4I)PLlp@U{cy| zUL*hSme3_2UH#Wd--I@_AaEtpI{TOVQMac{bd*vAs{D+@rCQ(uMsy;J^_raqmHfnN zhdZ6NqAx}dQXV`ue^x~KNUt=InzqgDgvo-^TY|4XEJ&aBIk*O~WRas8k1(MQydL5L zfJtY;T$7YtMv@XA4@z;@!wxn$wrJs8Y+?_3aDs(Vtz0K~J%(01f({wE{jU{bAm_>J zay}rY73q^&2 z{^YyfcxAm=H<-5RX3Y-0LYLZU3^vyYf;r{eVm!%WklJ4EjP?`Q0Eq%zqB^YZRSRCu zVE0tV`d8c0qLC+ci^&ee%G7Je7zW@vTf$X36g#V@R|J zM{N)M$H{8ry#8K~Gy8la7E2uP?6EJ9)lMebWiuBy08Nw%t|5m8Q_JLpC?4xIS?hab zu$#zs-|x||>HI8;!w53-y(b6$5_7j&COvEW{m2C6%7Z)u5zlw^lyX%rGjx9mHdsO? zyBI(aw;@A%Ex236J+j3;9}u5H!S%uFoX{1Bh9M7t%_e+T0|W)+`5;@QqfbJtI&Cht zW~!ITc}%M=-G9~|9w$SudM!l!{wf3L&?lJ~!m$$42P}J%SCKdyO7j5JFjF??_W=0l z15g6SlG?NE?UY#n0acK=4L~hrxAW;TCxUw%w=sPYU3Zq9&gH5{2 z0c*=d<%(X|W6bs86%IT85RhYRcIq=>tOoyq-%#{33lZPM(g+DeLr3q8{Kt`A;K#5w z(B#^e2eY(^AChYSYCctFu}~wzun0$kL(nysfEaM`4TeNSf9`C_5l-6@`HD?1aHaEk z*|jGbDQb=d0bcut`@78iAm~oWfWbO1*B;MRkYmFl&D^C1+=4=6ZZSEM-Vgf<6DfCo z=l32hOq6=dUosh9|Hi})!qY%@Iu`rPZIhs(N5*!_M{{cUT*>2+DYsi&&mNxOo$;9?zPzn0)BaAdU2MEr;N^=5fSD}F_`ZjUzy zn|0X-WRb>X?o-ZqU)xj)Lv`#emaiEc`wb^)G{Bfo&yTm`WIk*Yy%jTF$7KSd?0EV5 z3^;=JW0?8w1%VScbPVgL3)d%eV?s!sV+??xbWw)^o(OMJlAT z(w;|HRY8d%G%t)Y5@|(Mmzq8^gMs(O?2nn8pV(X0zi}8AFoYY8QSt%d5}&HOy|pUv z0esk?m^lx|P7>XW;*k~6MqA+}->>IuDZ;K&zOdxT8((Cb;O4KSk4Sl$h!V`88@tECX$dw8ff2n9nTNi z{Sn<+L>7K?Pya#2HO96SVs{AuFVDK~q}a@=L^?_Hi)7b>KrF}{&D_p2(OPE9nh##~ z3(^>>xu^T9O4&Bk`vzT}`uxaoYR*?Q!p?@)YPb?gHD}YG)*|@6`KrcHtNh0!uDX?e z<-F9K{OfqW&p*GXM&I>cQ}p%yj$RWN;f-Bnq)G)DN}P%D+!~VCPW1w!lr2MNB7UFLrU%?#S<0O`H_#1%-K~dudH6me`!&^(qs*P<^X^#;!R-1fWz2RxMPvzlvW$jSJ4 zCDMD0_rMwFHf&Q(-%{nh!kD9pr1?&X?415=O^b<7T_s|USG<7?;J&A!v`F0%ZE%~G zJq)t-{%c|eN(7V-iGW~85GgW)rV0pH!75A4?zRQ5;nTjJ;LQLFru0x4s(P2g_x($%=U%uzvYhlZ*`SDu;pM zZA`{=+iLc`phW--g-e0ckAhJA^nYLXf2VKCMXHouO=g~lWUGmJX9Pe!59ZJP3o?gv zrnX|zIc!BbePYPDLWQ1R`+DqV+#}Itwg6H;vS}K}(Jh6PB<>#5sdCbTwtN?|Ty92G z8kc>P_n_hjJ`bt`|Gv)q-f`?vbad6y1iIi3$;+F0b3F16 zWMoC2hLHu@N$eJ6a7Uj)FvI}b83P-?_l-6Cc;5AxV1otU`<=yfy7X&4{e)R9Ey?eO zMtNQ9LAh6s}mRY3EX^Sog<{D^}M#Wao6 zK-Q|zTINe*u~27Sg?>3whlRmj2nv|_?xrs^9$Jg;=M2rf$nL$vSiM16Pge3iSz${p zQ3plN+q7LbyyJOyY9I%h?15KhPklW(#HkBf{b17Q(xqRaE%Qy*H}C+;cmwc9gb@|_ zA?;<3{cqY1OK4;shsMJ|HDDR@YcWU6L0>#5TH*>|nYNc{*U)!Fo;tv1s*cz5fY{`h zU%aV@j=gBOn;#^S$f!`!L${#{MAD4olMPBd0~AVZ)XPB!9uEo=WsKI#d2JBzt_Gh` zT~W$TlR`y}`<+xR8>Ab4NvpqgKhK{rH-Y^R5^YwTUBQyhMj&=vvMpcx%b#&u)iU4+ zi9C!N&K)>(KjZCo`pCgKq2znP`{8 zoXE8M{bLjfZ7IzQ6<+mhRL7v3!2Hw{5%96yi3)#g*_I2sMBLd9L`eCb0VmZ=)rCY` z27TLc?s=y8c3r8Lbi1vX9kI^qa2&AW_o{>S{&1w-a4^b=A%lPor^5`uB*8EL%IsH> zTUMTCy*J!8y-1hx8*|G%mDLulcF(pKYr(TRcd~-&1{L}$QcqCyv~g}7D8)+iS8<0T zti8CK{;n|H> z`BiQAC2JlUlV&q`m){j#-hF6A;VYJcmJDBXOGBYf0l+DdBo~$JTd?mFeB=Oo|MP=D zH!uNDy7KXS!=CF|^hq39PL`Ssi|>s|?I9z}f!Oh7=XVr#vg zGcdi3(WZ|X{!0J1jwWg;qxSIHj*e@I!ykd8z={kZIjzy3c#acXti3x66Fe$DFZwyT z$U12(XS5v$kIFvqMe?>P{VQhH4xV%p)y)V0liv~d8%~wtXZDq%db-D>eD8kK&fa;| z|J4$SKD}J|Iw?a!$^n*hY{dm()P(MN9p6@C+ZyOfAx`)g>nsc7==nc%hi-v+3eYpA zt(D{s1*vfJ)-Deqi6b_wsSgMk61Iq?A98!v;X01T-H9@|N#7~1Xms9lJvOdAArOA? z0vDm3d0EJ{&7zUQu6;yE$7lRaW>d;$ftYpVYSfe!mr8aA8P_(n^+eJf$>0`g*`m%M zr{LE|H%8{EH`)#a5whQ7o>sKchGcZ~OI<#N2Uc(h=D)G;pfa?PeBh(t0Gs<Q5X_jfdube+K$^Y zZRq;muiJyud9@b^LsN`SEdYpI?Vm0$q!8}o(zd8%2mw=hPg+3_W}$S~;(WKxA~i~N z58U^Af53RmFdf#euQF|jMt0n60T+CXMc19Jd~E@GK8HAZT+(vc2)sgW(E=Y+mPv`nxB=T( zJ=;qc*6*2IA7k7!IxI~;FpZ9?P&Y(V==-frjmSO~7f2)9`98g<1*>tNzSfescoBy( zzQe(KeffH2@}S;&M~>A_G*94CAJn&PvVlyKeZIuQ=MI0o*3(io0~pc=U0a$oduHK> zy-dY@jOGz>3S}WBPstVY(wB=u8$2Y&w++ij$wNiQhon14?oT6kJoyV!O2bTq^Z>3+ zgA3xc7`y_3XDB4`FDa7t)svEAOec*pb{zIm*%oL_0Pn*(9~A{~*ZhNTmNFEf;-l$1 z1l`;S&(Nq&t7yZP+{+A$WANpriRh%NaBqIm{2Pts?qQ9F5{nSUSf$7d`B+ZXDHb!w zr*?chf8o3%vkXkt;W?YyFeo((LRl;GOE`oIt8Isbba46j>fZiR%qH?7KO2g5sOwzp3cGUL*h) z%HA-mwa|#m{;Nq21OjFaBp?w2c*>Y$vUt;*;P1N80luvJ?tc<=3qUS}N$$)xoHolo zO(C8QC_|sdqD8`E3ZEF`4Z=fn^FAg@-vV}~BtH}XK!Yf+E_vDV?fg@px(Wbv+5OJK zo8$#VOz%JNd{3Ye)+CU{XYnEyXh-J5TTnkHN4jtBO|SQ9f5#ewQug+&%tsCDOsFe` z{1q;c2VULG?;@E5(@%Z9i6+lIhGqeW3il7=wWzfqGW5UmH4uLgUJiWw0 z*H@RTuU?!{Kx(u@%QR9CW??s3>q+%+l%0?o)yHP%Vm5qcUyE0CQckU==(f@XI|JN_ zD?Hf3W#$Z^js)Ej8%IGQ;m{cbKLFK!GMA-Rc4!##^=!__95e$YU5nC1Kb+N)ywTNZ#uj6cyZ1LFVBrsmsJ<@tmBjSsU^z6jecBgTDY*r3aB(Qxg zJDYsI-@DLYHOTIJh|Hd!wydSSJ>cJ64>kv;mn}Gyn>C8XX)UMGANBj(j+TOZIXE<> zl%z>5%beEhyW;31Dd=g$`=q}Xz24Ckz^CS%ZSdmQ=xEQ0;+9dXimevrnWp>w!J-LwfQ<@c#6?4+;(_H5$PNL^&?-?p_e?YcW$M z{7{ZgVV>UmFw<_JcHgSS5WX^oimu6YPb2}*z?mUaJBp5;vQ%C0;M^(|B)Ly8kN@9t zv3NIpu~GSi4Z$y+L?9e{6h>%^@6zSLN1xgSeTN^P_xIL= zL1cO+JctV30RhPSJFJ&n4Y*sRaGrmZq2AATzjP6QX4O&*d;n3Kv3w+CG<`H#=tz8B zplLD~5BL0aM9je&*|CLsYuVKxpA(eW$6G)3xTc4`K<{Csb=rD^ZAQkZMA&s&w)^o` zyzO$VGGAMUV>&`q`^lYj;{|IQ@&FL$CfZQ|)l7=zJ_{v(GLk!Nd|BC>*-u(f9dI=# zJX2wyWiM&mV^{$Bx+;(x9$PiR5V9*WP?lY{MJ= z$oLcyOQ9mi+ET_E0NYe5DB~(rZU1(rW=_V z48AoAnf)#^nTCpvsk$t5O17+jq!Av%NwKXmNyz?zh)xW@*pX_Zohwf0lkIFtJhr0& ztzrqavFImr@+4#_f+T%RAT>?v`F}Cm%4=f`MVvGA{`B9)YUv=^WX3)xght98au1hD z2$lDN0}){bMWnK_>ESaS-F~@ldBG3iO^+6^wj_0`CNn$K@^ zBnmW&5JFtbBlr8>FE4@AKHsP%%oFHE*+)lh|DuQzuL&%-IO$mrLlPxIwLO*65J!Kj z(Q6A{EMiluucoq7nz|KG0~~gy!g3X#SH4F-Hn}n6FYyrTSL>1w#lA7<-5<%)!)2z> zMVF;JDVKF+pniG$w<4_h9E3(i=%>S*LSRO70<fPX7dPbv>)4HXq9!LP;qWh4@0;U9M624_c8OQ$J2r zII@ixUh_gZG9Wwtt#O|IURH&X5mokFS?NcxKB~-*nc}d8_jx~A6)?cB^L6Dfa8yZ? zUn>;Nhm)a&>&q&iOSbv03wJbj>^~tiLf_mC*%0omFJ_6_PX^iB-L9-Qc34r;YHz7i z{{!EuJTAWcwn}=}zsS>sFa)rk&XnVAZISS~NqqjRYys*}>O{73)g5;%+MR{saeo8$ zzDn&?QD(H0kRbPW_Z&hYw*PnmKuV?1Rl+bi>Bzw61v;Opkf^|Qh(%xJ%n;n5NMd*~Xcqx<&8my{JjccRg^DAgoQsM{0Khk1!XT#*EF zK@tmA388{Pi_)0PymbvKM6QMY*7*rxk-vJG=?l77H_ty>l`MS&=sUjn zNG<6hfY~e7R-5CQg5Ql#BcZ+|N%)rb>%?>eV%DYOWX}bnUc$k{w5va{Ack0}?VvRG zk--cVar3^K40<;S4hyL5uNCE4Egtvd+qPwBihSmQdJJXRTO>hCjX_n00g-o;awUsf zQhV0ge#A>$LqdP}0&v2t2f$&Ak+hys2Mt02igY}3hx!xRo%CV860j%H<S(M*vETjwj_%!hz{NLph0s{L5BKZHGD~xggvV?$QwG$E0 zQ(D@x;vAAIE2^kWCaQ5E&Vh$qDR_1EEM{yHM4l57%h5z-vREnlE+wDZGtMn^?UW%S zS(9^G_a80ltOdmI+FBqxNNCZx_c!9Ti;i`qI@hDs-D$7HA`V%M9-+QjYifC<>&D;f z#z$(4F;1Q1zkOR2v4k>59sqQ8xlr=EA!888!A;JMQ0OE?lGX3_B$^5!JeG-M6UL?1 z4l_VJ)V6u9Y^d$>pcc=I$z#ltYkh(4qc7i84%Qag|M{`IXKh0EZQypeK?n(oFZERa zGlU-jhrP_Cd2!VZ9?m?+qx_TkWlRbS)#_}NhuXX%kw%)(>(eWLKTdi3>8-)o>FFiN zsv9J};o>eo1ig4W*4QbW+sXBPm?>bKS zOg`+UMw8l*dJs|v?2E5|Tj38^dz1jx-nQw_M^Ks{yz4%AX_661LpdI@FKs=fPW9>s zX6Fm+1;+YadTcH0`)n;9(x0Rmj9X&RlkYi5%&nb6oc!<>Emi9RrW!*RqVExQN1vhQ z_{OKv#WW>ioQA9arKl_UwTaw~(gVS$J@P$7kC|@y3Ft1g z4%xy$da;MS@!sAJ7%uo`qtAq=lj1ZhResS#jMO}r6#X66FY=6X->Xw$one(r179Lf z{gz1QtDH&ewi4f7e?reTqCs_uM4&|pR_5HM6piW6$XyR^z#2pCXb7$pljOtAIk)ZX zZA<_BsrQ;M+gnFy;tLz)6areIVnu%5FaeE zsxtuZyO6n;oxGt0I*z$gx{HS^Uc8XIA^*+yvrkE5fuVT`yWyXCVSIQ&YKTI_WsnK- zXP0x6PuaE4?ydD()J&(J^v9x!%^}Awv8Y`}#QswK77V9Kc?@~z^*?jwx1{Nxd4iGC zv7QWjp6}=^byQ0XVClP^!&QGh*tE@x@K~Qz6| z`MP}NA^zLQQhOzALJ=1<(D#adAots9S;?!k8~w>iy$)N&DuBh7Hh4t;Ie2S@qZ3MSHagi^gHpehnQ4 z=Vbst6~wVpE|MVCqs>v`V$=&{P4VZA1ouBzp#UK2mMh+bvfp{v<{1&N6|0f^j2vMd z*qSqob~5id&edYkRskG)8Bb*0{5Q)!FTQxrB^5C8g$SbS_oDuHuJ_m(l%-e9¥wF&&ci=hQo=XhfP85;;?_Y z-m9pzh%>Wc{A#4<+tTguLv(qLhPS94f(O%Wu~nE?WGEz~JZitNKsYfB&qqi?gjF{C ze6EAV<1<$)!w?RP^f`?G2WB@vHh;;GZ(zU(=BO0fTR-!YiC2-d*~;(OB9>uz0PG;jhc;BWW@^4h9QB=(`ya`lwaAHsH;CJ8F} zH1p9C@!xV_K57Rm{$QuGhB7W@2iyZT;m1lJ;m8-h&MOPPo3MlC-4)QJ&vDn2FoGy;Kbd4CL$4A}Y11?;Lh z9-T1?yHL0c=4y_y4GJyJX%%`V6hGPcVO5^Uh3Uk|ca~iLSks}#^6GBILP_{8>mbI~ z@!N8s5q_EzJDoAs*0BL{DM?KDXNKHyQcf$;dlhY`7_Pj}cvPv|jxx6=0^UiQY&kS` zKifo;nI)Qs&A+Qw5XGa-e1F(rW*;EgSsq|A!R#iIx&!47BJsx#c-8pfVQVGuZrKPE zH^XADZ+|*#yh(5W9UyH{`7HUM(Kb;QUs(C5D@8-W(X+}K@3ucMy0zMaHJ#!rD0T5tJVe&70|m%U17B zYGjsT6v^lA#)RSh6X2rHbIJZ6B99Jl!I{ zSiGZ&911${8)SJK(;hh^=_&Whi50;f_Ke_<(##>Tv0{&zIh?v3lD67##@#^wBttyU z#)#uFnudwPZ^anN|8>ceO3?pJ)olnwRKF)I=-O?Xo@;m3k_h?&7^`BdD;)hw0L3BX zZtJqU$jG}i?wr0=J_<;^1QkUJav=azNq=nEDX|Zl*=5xYCH(I%v z#rg=fu)^KWXsB%0W1_s5Qb_QN5$J^A6?KEfqgr=+b=l`#dMfS@PBJ zoI6kraWGLwzdz5$TI9c0J=m5D4nIN98b}#& z7%yK|pR8Zuzi(3r*As}30!8}bO!>)gb-AdExKU*^9xOJb4YuyJyp^CyPwK`qQFJ)X z2j?1_9;;SyA4Vfm2Xg&L1q+g$sGgzgOnd4IyyZzF)v;?2Uakl~*BQ=m4XUd@FZDWG&HIhPURL7Fhf=~^%}hNRaC zv+f;a?Ag?!SAIe6zu^Cd8h#}9&?vM%lyAYPwzz~M+E+?aK|7!DRQZr}0h^2Jrx|4U zF6OJr*|o`&B{#?AuSvreUr)+S(Pq>!?i3oG;I}yO2>JV#eYC|M4!R}3iJ?E}QgwcV z0q`J|i?9HdX-eHwP--{y&4DG80AUW3VU4hW%3IHfTQB(23k z=bj)ml+MWf*--y79BA_^ZIJYHHxC(6Q8(BH@&^y``mECC^z;P`aYx2W;T{j2)c9(> z!?VvTKla~dFI5Z4MCT^+OzX^IdE-FK(Cl|BQ9sjKgRe~;Z|^Hhp7TBwzkJcUmD%cf_I{bgU6=oNCX&rZW@dta(Yw-5PkDonjvEh%){0|eXt>vh zRZlqQ)LX&uA(2CPD*}RQk8wElN^U`RMbFFqIb?92Md|E=ia{A(>&sGL6|ATEw&DH@ zBV-+G-W90p%(V;dJYwxG?Zoyv7$H#RWw`O>Bj_>_qKKD!fnJ0I@?*oF%$4$_V*7U> zh@TmK+87uT8yFWd!0UVdf-7 zi6T!9O#b`iB(Oyv|J#-DXf>)>0AFFPL~ZvnoxqOdM$0+f^|D86Rm41djqvdMa#DUU zc5aMW;WWyg&$h2Ef9iy9mI*?M0g*8hIfMqh2H{V34ib)3}y#@ej?!lDZ%j)o0V|K%zLuLeqg51 ze8D@h6@)1F7m(2rzf+sx=y1s7MBOGK7pPj!>|V6oPw$6L*gB7&=CSgjefYyYPVG@? z-1_GE8ONh};rUZkdG>W-VQ);b7H{8!yiarI^o0O33Zk&z z4qB>&7Q`22cv-%njP13JlZJ$v7!d)~4XaYDUu4XX`xs;W#02)Y?t)5y?ZZqh z%HYwHgsL8%VHKr~1Q`eNpXI6F?}+MFr8==k0@4=A^wiysru0#|P7F@$E>>ulUQDuD z6zT@?uuCznTorjIlx)O+XKwxi?}*e*zA$U z+$*SzI6BfPjoDcpE%}k9vzkL7OJK5IW&A1yK!vv<7kr%W?rKZ3Q_p4vAV?vnz?sH}9 zT10($A!B=w|CW(q?)nyXFlC}L?%?gL$LjEY)RvHQ-Ihc?6+c@Xb2Jh8WUzqo@j z>-66Rwp%3kl{7Se=;Iep@Gw%No!Vg0K& zdrt9!1@iILmy&`vPt{DdEusV(J4S%?%7;1F91P9ldN=6NC!|^iVLLk#93{ZG=G>vD z!JE^&_Sg*K(26Fdqy~FAwX+|eV#NTDmM=2rmn5+7A59YVb>AC#>^PEgv z7A}j>PM)gEwDnwdDfGX`ctgfm9uu(<8U;$NZD zFZ(;Tst5ZDo_rkpQ~&AI^`qDQld36TPR^Y*yi%2azTHk+zhhrulV2L}M+LK^CcJpK zF!O9A#h}y|J?VsQ9(FDTU9s5MtcJ73zZt0iDUDww4Jsl{J`CqC>gu0bnRziV>Z2Oq zcv?8(^?Z-g|K5IKXbVIv9dC5_pDJN)@q%q6Hs`m5`5)BtUT~SK0vCqn^!@@o9TxE> z^V$;qt}`0WSJ{jb>Uk2vqB(N9_yqw`T{#4QGNu{AGf`E+C*O+*UMPeROktHMdqhTU zN;6c^1)_3KIQTd0F7t2cX1<~iHK32c=kI%_+ZL%cpaf}r1>zb zQ_GSa;|f`naz}a2N^>_!7{Cb9SDifg(V?!M!vWBlvQ;>m^GP-tL4Pd5)EaRPi!huq z0v-8Ybx=#ZX!!C)0wz?RK!~%xFANX1(LVzO*Be0iz2vVxi8?@jFh7kp56~KYk7IH7 z12mMZ_E1fp?V7H1YMo0_b3XDoth33}@XKcGqaF6dMH(6#ug>jactDM-G;V{3F#Gx{ zji(7f=S(B0FHh_gi!I&JUtAv!8`eNj1uixj@Sd_+ydn+9!%3iur(xwQ^pv@i*hM%Z z9$bI50s$>*mQ`YRMsB}i#((p)miqM7et=k0tihG=03ifi_=otvlvm-z-KXVF5SIyz zQSkKRESv~qg7#ECyqkxQ>3ik9TLNqGWRY89!n;8A#;t(#Wt;A+o=e`x}I!=<8BvbM2;t^ph;5Ym2QUagr@uk}arZr8GPHh4{Ww!EL%4p^1Wg$VBN^TI}RRCaS%3|$$7Am&I;e&Q(Z zboWX2McnRr?%pi5-F|hEd(Q=1{meWu9!mqgbS~V#_qpx{S4j!MclRt)9wL7BJY_f_ z2MOdJv2~2D^*p2p;P~?*kQ*RGgxXZR#sXTAdk)ut%n|)ELGfX8Bl}U&z6Z7M*@9=CE8{P28cyLqx=U~?asBV zGiLRk`3nwdhLZdwRV1ETRZ>-vT*C$#`>xc>MZJ)RsqWP4lSU%R5@s$47nsl;@3PdelfN@x<3Rz-`WQU;!LZHWb^!JyLTsH4 zfgVv03xxQr1*oumds;$}oh_W@8+Hu)^(1Q_j~;f(`+NaW|Gj@}X-eghN2k59Ge zd)c%xD*>>u#oQpDF7c-d_(0S^`ul|+`e`^n1Sh(I3kP9w)#fLM5%`nwwEoEV3ngIC zXPMW>8I>Bx)Zqh-vGl@$dT|7y^&KF9v)VvQA}N>15eGx=&rQ7c>aJTHo;}&p1vS1p zw!a!V+IGiXx*eB8!rBn%lRxKWWgf8B`a~UfDo-pNM3`H`6oS6B61f!dhHr{@9^rDh*AW$P zP;3;a`SrbcvG3s531LK@F!m028YTW0jEu)k1@NZ{T0-9{@;Dyh* zt;3F(<(Xvc{TIDL0!B3`2Gk@EQHy>Vua`(^QfYXQWlj4brVTLsg-*cr9@uY$xHss0 zXH6+_v0*3gAbUN(=f=56S$&6F&CtW+#2Iiv43B?i2<;vK=U%-=E)A+8HgBI<&GL>o zp8@@D;QYY3u%hY`(8bbl-WQ!L%~9qUd72bKh$n>40YOu(<39u9et=)`<+$;n%!ghT zo8q!krb=$fQlIBPKh|>5VURQ%j{b3b3Ax}Vf!uf5T?9x`k`aiz>FwFQ_|SlU;llbE z>kty;#K5fsUM&ie8A12kG)XGx$+U6(7L0pS!`ca5fJ8I)`@#IrV$6^S1WLF1(|z=7 z90wFn`g>#0O0?*5 zEr#FgXkf9l&i>Wo0`@xFtK;LZPQrZYiMk zoU0O{0VQt`t6OALyKLoZiBW-0k~awjy*F45|L6(` z?tC0^t6$Su2P+@eLqk|H!|wPs9w8U3njX4I-Ibbiiui(9@=01!6Vo7w07u0RFaWuYInkwa| zYG|e){{%7!0b}o8-h%vmO&jD~F+0cRoVWF`w<-tP-V|fYp&ESQ3GgAxF`kHj*%@O$0o= z&t_5_P_uMi(O4fWt;5Z$UiTH0aSaF2*G{rPr58Nv@)kl)T7KRGTX{M;G{1J5M)OpI zXqWKjXj%+@Y!oHbiJaUICfsr^I>S$mUe=w>xMAvcJVsXFUKUt>Vip1-+%y?S&jJy} zn_=}KTISe)Ck%PCRs+H1NlVm}HF;~c{g>|dS)wm&H<`TIe&&Dj z#maSlAVk(sF9sM< zVQ@X+vYo=}+-d0aCx&!zHZv*d7+*%>j_sBp{S5r9dy2-!IO_H1{ ziS#MI8fT~B{yhI}%Yb>jj-3Q<1GYi#Vr0bYSn2qwy$0VQ{)7-RN_h7cIR?J^jx0=i8V@j*0m&LDGBLe$d5;X+URlPU$(EC;g zBbJAA#WB9V0f&=pBqRn~@u)}-+Rt<+c>(iOPy_nEu1SJJ*^ReYMJJ#$UsadSF(D2} zdI}9ZvJTz=6qacm8q;W=$q#8f@+!xALjZB_lwky*jow$I(Rp`XZ?5)5 z%ImoTW}oFNd(DTQ4=pJ}DPponGs7a;4CQTw`7iUvSRDoqM&@{+`hbJ!%U{AB;txA%yD1UpMw8ByJB`*1z^T zMu64Y!(@;71DKB?le^E3-Dv7PS1TRkS^n@)I&0S42m=ZrG(l7e%?DFeg-t-shef*6 znM({|?_e~`N^zv%av+NmqYzY#=#Na8X8pbDDlp8wK?2!Gly~s&28b>P$XOQ&d7Fb$ z$%;RUPVWyBSSMH?{0p8e*tbQ>L}jixW-LjXp;fOEJ*?N%I1#p8mxic5z@Dzx&INO1 z;iPy-GJ{X0b6d1RoDbK(-m|Bqa9$2RaFgdTJFFgfdW^eT?)`ck^1IhM^W`KFiMVAv zm(S@t-PInHJY8+y+(Ts>H6U#!y>V4mlxo2Z@3Pr=bi zH!f}SZ7%Q3ZVL4lCnTyuE?s`Cojzoij&ws7Dva6nn*rV?_@h$XgLUoVe#dRdJ5$FT zX~Jz%Vo?^zn{?7SRl5N z$g9HU|I9Esd@kQV;{ViE!{z2yBG{#6SUMa%Z$7MJH|T!+nDSm$OafQY_#+C^v|QH_ zCTj-KwDKSsH6!P&;(9F0?p=>Wy7S@Pgv~HdmM*qkE7i3wiGgvcrh@0!!J$V(bG+E> z9B?Q&UhcACkpV8aLYfz8fY1ghuP@ZNB?}bK8x-!`_Z_Z%J4FtDXJ7U91(a*$2#EG( z46d2z{|QZ-{43#hZ{W_F%O$|?>`PU%QgR0Sepmr;I=~^IvEMZfj62KN?JM`=@=E6H znxz7lgU_KZixH)@9i=mNSrVceGCt|P+x}fPoxfTzHf%r1c1zrZw>&DhSayxdJ_&r= z{AH}>Om}+V?!S2^5}Kq}t2JMuE#qr7>}{lm{=7h$*E_i}uF?eq{{$rPT~PfYk)n~^ zS+`fZ^93x|4j|?o)H72%@lD>P7Pa1Pm7$CoEm`XT!G82nph8!MShCl`PJee!k|+BDp>}}I!!)z&c&m!JB^k0y^Uzr z*>^x?$%>LMblXTS<$7v~#?i37S&6@lI`LzbB^1u=qU}C*xIPc65Cy>7b4>4Pxx)W7 zgee&+ZNcN&=yRhac}9{)zFfVG7bsvMrVx*rtTj}w8`I%H57>>i<;aj->r))1nvVnfn=X zlw7I8NPa!<_a^3NV{5cV`E5?>;f1HLSI#<740%KY2 zDjr@*$pMvjS$^VL>9do>8hXGlm!Cw$B>tslVp#d_Do4f_mv|y(ah#56VatqN#77F} zwB|UCFNq=Ud{>K1TwA4t6*jEKHPl)R!sv)lxJJh8fns z@?}`DM`VfkjB2cKBA;Tu6VA{RJr>n(#0P3?Y}wWny*x{`3X`Po>tUP==PFshHr;?G zTWKZ*b3R&=w&2%IO1+l%n6@GZAk;l5LVuTA(c%zF)oEKBD9hadMU+$Ga|cPWWbd ztg1>AS!_UF#O;g}OF^i3Ger$ZoCG;8B4L1CQ$0t^93OpXry8cT7cCruN3xFckSJqL z;4qpGt=Ii7Ys;SIRwr>n`Mym@tsI9?!|B0T{ zEZE&&bm2Q(gEvEZN|UqR8dRv5E2}`uL$@5R>_z9gHKc|Ulqg7CAP}t7)jG9ApGi=J z2p5}}->o?3@Ofk}coLQac6nI`!n0I)o7gfqVvX;JfQc+xPIzc>*nS%Vpe zfiT^z0%>a|+{?@u*LT7TjZ)ri>%ba4xi?4ghXukYsBOatz8^xUK5bXGu`Y#VWspj!-F2bsf4qdtK zct#|c(IE!V&<<9OIt7V>do)wMQ`Q}XN5Q7zX)vLvSk2X7Ri{T?U`FrnGaPXe&5D*b zgPE0onrUaoKqrenOD=%Gt#vy{{PiRxKD6!@t2$lOQ`@wtgS zZ#?*-$L2;xWdkEP*Jn@2|0Z|ns>X5id25nDOzRO)#~^NwEXTH|o!nhUlxm7D4-WX! z2m3xj<+lpdy2jNh8rw6O+YATfmmRdSrli)=sE$1{OQl)!wlFsOc|2gu z?W+vTamG3TwRFr1RLfZWer)_>13{*e!Okx4F50>ZdC?z8WT0)>g{QaTK1fnN?znS# z)dYom#x`L;dh~IVaNb)pQOKQZx2Ga!(Mjfv68IBsr~f-(j3{yLp_x)#-mh{Z3IEuq zJuN=uz3bWrRy?5^t<0^joGH+3U>tj^1LSsIv!T~d$3F=(zn|Gx{1^_|ebOUsq7_Yr z=dpS9_w;JT#{-_Y_470eeGte^VduFT5P%)V;q#jw!Dhm^M)v<0i2FdF)vc_AC1d6i zLE6sMCk{KN7$2t*KTz)<8+#D@4nDE`nx~7K3Tx5AoKAMVnBUX{v&s7E5q`cPJ=?YQ z31H_1NfogB_VI*wtsVc&u=c*fU53|jr8BQbFl4ZW&<;Suun5v`=4u(Y8aZU%0;gtg zQH-2L1sAdFFMAIv?!%ODj@mBAS)FxiYzuOpV}NVvfIE0oV=!NVW=^V>y2J`k8URt8PMd#FIkO z2il<^{F-&v*f>u}^I=0T_c4#z_q{=xk!X+5-JS%FBb{WnL~{!MMYF*NYbMrk+!2?_ zonx{ND2efg8lG|_DDyK;=`gG-_R7&*obH8Tn+>0HPBXP-$ZmVg5}$9EkNTU8CI9Ym z!z&w|>8fK{Q3+i|XGDbzM3shEjev!pKIze8lc@~oRUdp1$m|zRFLM(g`E#@}S)sRP+0l666`$UF?URWscz(!t# z#``M)IvQ?iv*XJn`IS_G&gIZ^TKRW_dW$nn>-0PwM%IuGxkDEzHmLHDW6GW#jlv>8 zE^{BB{Oiu?#M(nu#&R;Or(6%#BHB^de_=2&nq)f<^01SYmqu9m?dj&JQobVprw?`aIs#f7ZJ+ zjP6h`dny6L;?_xMn?rd7nSdp=#q*gk&)S69XNYc zCS2BOFHX);f04*-)$q63TSk!%eHZyN5RdDC zI|3M2I>kp%?{C0B5O*)deW(?8wCGz6d9 zGDdnWD|h^}ozBnwsf5{@s5rAM564%ZUt}k4j`<+@bdNa4YL*SF;j*>py=GJKZ`~3b z4Yp|F&|53y(5Hv95Slz3wm%m%y-xSYzMqgZA=6K*K|HMksk^^H~|_;4jgAH44?6jSi`|upZ+5Ta*IM$Ef&cyj;~B z!u>YkchgGbQr_4ih>fMdrDljfgVUu(E+=Trzz;)F#TY-u8hNvVDC{%`r zzVbE(O~R0c$dPNGB4P1>9*v}5(#G3HW8h)(Z(vVBn)1JmZ*X_)6eY)1U=$Rkdh+ch zEZk=$dJbEFt1;sp8mJw-|q|-)NfMXWB|3bk>esL&e@yNOY_}a&ybywB-6>JdoYO(#2R`CX7U9?*=epR#hoQD#ehaiQj~ zvxJ)nC2Pnq_d(5DiJ2;0qu%z3cZfR1Zoa~Zx4h3M?aMTqlQuk%^aT|tU4~TQ(PyIg z+fiSXZc6j(QvRpM4--GxtdSJ93B|`X3!lS!t*hG0Q5!1 z1NT>%T>k^E`KHOKA&-eSW_VZyf4oryeO^%nyvB!{UN>6eeh@e6S`SJg&m0Tm>~`e` z%TDHUJEG2^PrJ29VFT)KgkDPqSw$^bDU=fp?2wNEln4st2g>U%=MGV3nm0*!Q+2se zj`goTU8{uV=lQW#+eF92j8C9mM{jv0wM!cZSKzIX|Bi7PnU@uCmeQK9vZzg+t$7X2D9OGdVbh^lJ+i`nmJ&PPwUP)A)WX6}x-7TjB^z^C|M4s){J%!*TQKFmoS zO*sp1-r670kWy?b#Cho#_c8qJKQncJRMXP2?tPkp3s5v`&OLCps3mNvLd0Ous@B?X zJ$kWWx5BzfNbaS075(u>C%koIuY2~RGvY=4FYm2V09?KRnHy0=rlovzmYcD~`7vhG`nk%amJG zFk-ndglW2HF8C)lx{UCjC`^VMby?%v5hJlW^YhxYimDjWzNGRfi%|L)Gl1ABp*#LG zMhexBGQa!cznHU2>qB(Lspa)MSv8r_H0kY|Qhf~EFgEVY%!7o+|e?)-o%4J_FtejreXW_^& zh-qFpnU?O&?lb?h`(pUcFXLpuM&ri=dXW{ZQ3J^8h*(2a`kMc}wrN-^b?d$V?NW{q#Cqbg1q= zRQXI=1k)8!`e?6GZoxYK{L_DGMUP^tKK=gBN>F&x1m9hkC zoey2>usW)9P&{avvD)!wgNYqtW}e3d(oJqs_3p@{<9$ygiT>B|_`D2t2AbF)7l`Tu zmmo~MhgX53L;5#z`2+gCtuGtRR>#PAl_(iv2I7Vlgv*;20ZRibWV zZt0B*jIa4j6VxBs+}HNxdqOD1No2Ay(g`>qB+qu4_u2Gh#*5RNGE?7MKd@rcUk?=9 zjH0Y4%509l50n4z%4X{}EoUez{;huYC+`1VqOj(h`@ z(=9F!NvgZFKigAM?j_A=T2qF({e3Y=Aa*6aZ=0P~yokR*R48=9@d)R*YZ@9zCz#>4 zE|<3rAOv2h)V2l)%n0tLeg3 zvMiq>T@sCFBfc{?2*qPLAFuMhYrdjVK?PPdI1%5C7kp-$RnpYsI@?VC*^0aneGp{a zZ2pDXMh)@G8GRiaSQwP0=(ZK&yUu^>$E%@pc`UPT6?}2WdmcN2HDVYQY+t9>tn8uv zbu-~t?0+J2z%ah#d~5rWXW~)$Fv|V}L>mvEgRdElTadMPzS{C{Q&G&(}>D(I}YCO zECwFQ5b3r12Um?mHG88-Oc4`1WqbW<066iKf_UtZkGFyqPu5Azb`Bx_hA#*&EJA}% zvzeqoTMRWkm&%5w(}Y8u5RhB*|HP0pAS=xwi?L(|d{vP) zomqZc6bTVvU7_R2Xda?xgt7Yaz5X7ED!47t{SK9F4vvcECr=fK_&c(|zp!Wye99g| zkf;Sm2=v@$cn4ayk~(~(Ir|LkeZA2T+&Z843s_)sRd5FnS*|w5BdQPc$!0HQfj3g# zH+TwBF&!~&nwUJ4`09krI86LSV)k7%(ipS$U`qc6dp)AA?b3-Y;kj7iyR}j3gV~s0 z)(PkD-&+xZ)*c5pe*6|wIUOy^6X39Q4oJN0J)hcP&Y#~8#Ye*dOf4J@1J*A5zzze9 z<22qJ<9r1Bhk+}NKjKkcS%-Cl;vp;qQG7c&b}eMNgkU&Yf!CFb&wOW8ifnRWFBQ5U zOSUFtX3~vzC>b~^r-^>_;zn98E&Ntxf{=~yg)x;<>LZ-5oYAut>a8c(A0kYC9Ad?8 zqh~=d{PHQsK^np=ZE*aRAFdQjutSV^E{5`5LiAgq9Lsx|D z%LB?5_cPXSCa;FgqHT-%-bI|v89Hpm#5kj7ns9(&`JZR;1Aw2`JL}~w1T$@UDc87Ux$~!oz%NMbcmeHo@^#_J zK3|ev@BmCJT^5rstoW>R_pyVvd)VjlXrYr;ZxJlw%5Do&n58=L5|5>YXwMvi8i7ty zw*{OedRh(s52*=E#sA?9P3urvCAephBZ&IaGUO0vlt}*;_iv1KOyePcZw%pMqgo&U z38W6X%SKGu(zy&j3uHt5U3p0|aS$}v_W()ww?ih&e_qU2aRcnPgqxP?BYhqN?+^sb zi*ec~rgw{DwA6x0A-sZ(O)7ufPQ6`3VvgCUjH5Ki z(i)p#(&hjFval~U2FJrn^GsdqG%lKDIga|S$Vg$LthN3t=&M*vuz6qCn@@$o1)e5K za?dyY*h>Is5&mVQ*JA{1+{dX5`HPgsy{^GKBZcZMu(w!ELLXub%zp2nC$SQ}8&$Q0 zgJ3LgI;9#5r0aL;h&fi@Y!grZT$pl@@csZ5AT)I2&le2F{&h0w6$ByvZ_WxhQPOBK zca;43WPu%OYm1MMwrV$!d4E?L1;aLCfK7}to2bsg2Z>Ls=sOw$&V5Gi02)Ge^CSEX zIl~-AqOFb4Hm!dfO@yu9wpC1jm}qAAKJMKn=DQV<825PV?2lkm|f#8q@s2&n;{I-(H9_*=IY)`;cw?hcTEJ%FKJ_bTr+zs1&!{&$(inK+!c7G6?j#_PgI{o0LIqP`D#5J_Q9e>JW#ZsOu8iW?k)OG9%F!=09cnuelx72O*Jxx)uG%M_8m9k7Hd}K zawPHXzo8XHZ;M3(3#&-MHek8dz}wwK@#8}O#-;|abStwa^BCPgu$dQOpb?LEyfivt zumh+jZ213ZC0NGiZZA^Bb+LvJXx!6Wf$*IRgkZS?rXN*%r^Poh0TgNg%m33K8Ma)psNkB80EcKbmxaHUi7{K02zMz%` z3^{!znFsqe_feYUMejaX(_JmipS&TvMgYrC=|Nx zbM$R0KQrn0yG}n;c#NPR-plyQsQUHJr+Xw6=`cz}|4vkf-cQEkEdMh`x9 zyD;NEIZrf=%FmFp8Gy_1ZyStef=xSPrT$krW8l{B5mcwVSQVTzZ1q|435 z1aDGN{N&0C;s3g_Lj}qI41W#E@*=jMm8#)F$>5NoANK5{fXE@F7Fo!+;5SGZmp1<$ zsM7!N<<^OB`2q7p)=PXK?Hy*WmtAker<5Jlh564Z?;ZvVgPMrIVhuhbJd%Kea!w<7 z6xr_aCwIh-u=l`;KxKOl@^H%XyXX1-+r6wgFiWFM?)&5i#>W4-LSF15rqiNa5EAiT zA8DQ^`)dv#o`CA$G&6^MPLSon)I>*xEF2Xp6~{Sz(9c^7fiQ@CpV|%K`x%ikxJ&bB zjSr?TW7T4pnuT*5byS$c#LULB-&B_-6#Tq@cN+p&W6Qqxv7VnsewbS!KI%Imnfx$ENe-n>ZM$2xJLFljRhXO|x=^RADDTj0Lo4VMiTF-*=&n%Y7fRf$s z*7mp7EXn-x0w!6O`W8LV)@2Nn z{Gxvz@{q`huuWL*6~hN{PI8Be<1Dc8W_pY27x&e-vetfn%xc7}Y-t*QdVaxokcpFf z#z3ka!Qd6~^%>S8mVr!P=Ry3;sflV1CujB-*$K)}bAfo-gkUzAuTVYi$G9&%$9{4d zMhlHN#!;i$_Snq#1d3~IVi^~$2aBX^pN=t6S z=UfJuT6sm)y&}_}p!DmxA@N?=RboJ|{Q&TYBtAH=5asCJ_@#@zi6yc-R;vKp zLv>lZK6!s@Ky_U$bNTGRa&hx-+R_!AfSbC;3K`~aP9KbuAzyC@hDx`G5g_5K505K9 z<;Vx*FF0nrUi=orz5k9~zDyTG&q_p+Y!cD1j2-E}v-Y%rN0f~rb&)2)GA3& zAQy2^$&&j0)m&dDvCZfRGD-%DotDDRBwoSZFe0&3QO%IJ_F|*bDx{Gl7VbAeD8(bO zL-@BO+k^@dP9&bnvf|3ag9eGG<JVgA9BwDNnm ziR82%rhfW;n;^d&N&=OsYm(qZk_F5c6*d5aPG2RH-q;$3>2s1U%=P}L3<%aP0qS_i zc*HDdaWJb&(#O6rCvJ{I#(6kl>*}bgC`ZBTJLaY@9643q;D5BxpkuLCr7p`d}GJQ@1NiANcidGD2;b~YHWNTgymRR0VHmC%!?n}NB{ zd_Bb!WohSXrtoRn7PupH=j_z6oE(OpvLqkbi%Xv ze41EypE8u0xCP+#{bB4~hLBQsn@qfkc7kZ@r99sdzBz~|y+a0EKd;Zn_vPDW&^v?d zYyufzmzfNS73K23|J!|FD!YQpapY-c?=*d-zfGmf(k|=1{S=j(Z^cB~dh@cQlC)RQgcQ&=H6=b= z939dTx_)KpzGqSEDlEr~&(;1XIBF_F-!TJupfuf^$h89$g$y~qK7gz@`o5M>%NntO zyvL6~@DqN=i+}$?OCy%u>XsH=eRts(9n{h?->IB*FLuW|0iJ#9abyg)CyzZ#H&6xn zxOTx^d8RZZAmC7{*V81~WwJi5xTwf0@!QHTLymy4D1ewkLF4J1=a+=Jq8_pT)kjlg zne(u8(b9v~7k>SJ*n97&rk<_wTR^3YG!ZEQq$6EKq(ndvL`^UBC7F<5}yw&OeZu?3pu}b7uD5 zpS@>UavBN8h|#HVZs&a&OpTyrYk3fe^|E+HwsD%*$hU{fx2fxLYvwJVY(@OI?Gy-c zyEb-Rdh+Sf)w`cuke|nho2yXYTty!*Fzw7aI4xcAHIFzpNl!<_%Ny0W@EkBl%oF56 zz-5oDo%0ts`RgCw*g&IwHoy0MUQ4wi|5q-6{eH3ST}H^Ks*e-*cHPZ?j(c9Z;P}3w zlL!T~vTyIeftHQwiVVet=&RB$s}%A9hg!b^1ATub-TgG_*79I;Cs#FWN|h3_>DGd| zfo4F@T%iy}1ntMr4F=&lC0p)7{8=buf4$GfEZl?C^`;Zx0+&kEG|Bn$F= zBX`AaSa0cWdTRxf!^tGhwemf`S9u?ZI<;2&Io2OP*%$g^vqzbEx>=xwX;sCRR1rnG zXwM0&--h)NK?C z=nsB8_WODk&0T-9f%GzIEuc~aB84) z=Z{LdIuOjW=__mA_NNPPJscTuQ>I%OM4-NKJ^#+@w%O!)+4S<($0?gvgr3C&+g+=I zjgRg4H}wun0i}TKu#$MaghbmRlf2Cm<@c<)bcx3c)hFN5s!fo2+dBy=>z}I8U0ZRg z*(tjC*eB7yg)%GxnmNFeLRzph&}Ihc$;JI@HX>#@F%Qdc$zku6vJ}aDvit;-qH}m&eTcZS>`a;{gh6i|fbV{kfPAveVPP zGg~t+eKP~EglU0PZbpu*;~PbMX(32;KbXc=f12Pr7x3j(mEEP@mW~;p=^LAoE?Y}& z@gZ-inN>VNK7)qt+qLJ5ugGYziD`K#DOss}R_+ebmn7r1dG{kBMxkzKUlu@iJ`UP& z4qiz*Zjty5+!-LBcX9y^iE&3`7k6fjRgxISgkwtj1ZHPYXU~IM@S`txp7H5u7-H_q ziByuXNF1bA8epac)hpK!KXA63`qw={AWL1zs-UX8n6Dh4;$lkRn0=-iehH0-s=eA9 zu!~_lSD3^voBi|!wNiP8qj#yH#QSKS>>IB6gBj#x(Rwz1!d?zmWA3=NB@g-J%C`Mk zkSxB&DEE7-d6iqns>Z5a%}V>7o}m6STQXseX53mPTA%>`Y@)|C;rC^3V>klA3J5;9 z(gIS??etd;-=uczH*fyTthFsV*JyhH>MCYqhHV%AKY?*`yDyGv;8on+y zsu`f3VXRT8IkjD!59l~d$tyL2<~JdplNVqUy!|2@x8--9X@Ldw9;~gD*9E4H|M0`P7g)U)xYao;P|{YzyARR*mWK7+|#dF zV!kol2iGvO=Z#;jy)k&)tjteNq5Y?HXMdxRL@^sMnHxCMTx}F*y!{PCzyudpTR@y< z^BaK2xt)aX$<)t(KwJ6_Wy^cZoBL8Yu+w<}!Fw9N9m8d%`QCKilNJ;ix7SarIs-_) z{9~H+-g87-sCjs6PXS$K;K*JyAGsAa6Ff;g?BIQT7Ca=IUVe&a;88q>{1|bc4cK`6!R-&X-!7T&*S($rI+|@A zFPS4=++IR;R~!G!X=)+Poi|GR!eB^G$FvBT=Kt`e{Mq`4ZC?A@ZZt<|E1ZQCML0(@ z0>B_!0=L@+K8LmEKu+c;!T4}Vuc{}Fi2|jO))yC+GhkhokdEov9_|e=k{tGl`HkXC zuHQ(;^jy2Thr6wX@8IF_ZWyE6*rwSwY*=N-vcn4&`0iVwW_-HpYt75oe|rxr+ZPWe z=7+r}F-!3re31R-f$HGCdbP%Itk&*L1m%|zenK{Hutob-pSu7JEAQ=d8T}FFQF_iTp zV&45&-j3jMgswRfdf<0df1?dF?S)aA_YjacSO@;J{0Sk46xnb@3C|g|1)~1pzqv+Q zL!9|bxh145kO6M}UcE(a2E)TeqEA!-UkyO)##;w5!<+ijdL=hoS(Z+|QuiA6-g~WY z{E&@1cw0>fYoZI3E3(#8g2dvtuOHS6Pa-RUljZ}dr{BxZ2zk!8@qwqRv^y-zT2M;Z zN3fq8mHNdjW+$L2^5hGg@dja~&Uj1OJ8bwCnpk=4UH@DnY4p>B4JuiA=GE6U>Gf2y zw&`OMX+f53ZpkXXGv??kVbpJjYO!qut9tBg{|w*MllMVO*ujkOS%k)&fbuB%FyJD_qwr(TT;0=JKz~n8&}b z{yK2ZM+qqY5{11#ifO4%xKNR_U0zZB+u0R~Bvr);3P-EVINa;3aN7VmBW-7RpG4z_WN%>zoHN0_*oX zI%tIDl}JH+L;|+sb$*L|aYW-;6=I}6Y0?dq+rkc9vi+#V{EP*G+-7@iv>lHXKjL^m zz=8SpH?ggMB2q zI1wD-K)6uOljaY=80qJ=N@C;D>;#}BQ^$2!@i}nTY6d^cbg3T3>_zxnlm-ALFxY}F`sevM?4D0d&X zu4PQhj)K=04{-^a8*)~RkW{t~++#y7!jD^yD~|xD0mt9;W^l3DU3r3zr4?zX_F_)y zvpL&_0^|3Yj!6S-BsnBeB;c0IT!N7kXpE_Mg91L&q$UBaCjKht|Dx{U5?YN;(%Lm#=n;sLJIcQkzsWxo4KLXw5}*f;~hJ zmuxI8C2m@D!G=~rl5uu(_++>zoCT`Jt@;&=gn4@j{xXoU>gxpLX6y&lYtU*NauYP+ zwsT#URiC;hc%4;2X>ta-MPwFae#Vc}m0<#>lt<8OOjSIQC{#*ynVZoC{wn$`GycL4 z>K6w0Lv72zD4N@sKsfze=9k9f!a2}ui~>D6(Af!7g4iR~FT2sE1ll`A($+L{`>O_2gbUYDY==DP4{hMJ*H?uWKrefK1ypqKT)rDu`urbB~+GKD4 zhu=#JsfMf7k45@33aWzI6~A11+5wf1vvOkOtB0ytlZ4Jcy79(JAo!N!#gWYNPSD-o z8S!eFC;8)XN=I^G%6FQnxBCwumCFOp(=v3#PGn=hV=>1NKWXN#Uok-?l>lF{ zTaY$WxjF<~R*%j-^+b+=2*DG-&xwPB?6OKJp7s5;pJq{}cIs__*{P`(lAD;P=bs6f z5D0}}P99Exq^&&-hwD2Z-MtunXW9D5@yG;+b$_30xR?o)Juy2es{j&SfPM)c!vh|h zzdRMMDI98U;b(_{pc5x1j*7agu@p4-ztgGlmtPvLSV6Hws_QOi1ym~;t$gh;!u+t+ zY-!p4q;r_chNMvQZ+$Z&ip(n)9JpZcwX}58Pv43qMOYl6?-`&fJ6u|v=6;+*)Ib5nw^s;=|qhlyw_!D%cpI(w@sp)B@9nNY>$LKTRL=-Ia%6m&!`a2PN(UyhW zvnam|eO2+CvNVVk*)2Sij*QhEnEtIAeI_pP+ zOhJqAS3U*H^ACAOpO;pR<51j5rwl=%O4{p=CvUoD=d8{OzbZ5(<2{8S`X!jvs^u@l zp+aAMdUL9v3c{&>{G5S~fsh2}53{nng-gTPN=*&YEifi1CK?nMmvb6O>eUtt zc=kGsyf&QX6I)6xD--ScMQd2;q|I5Eb^WSQV=s+@y>og0-y;7W%k2MQ5(zpLK0e&k zUPbL+k9Ra?vV8B>bx%7hWj{$@?PtHu)t-u9Ri+{$-uf!~_nuzSso-zDHyxXnEF-jh z^?i78T9@VnZ`uZzHK}D(GucsSp`=xzi6%(+3*njsP^#ifEcuu5#qs{xGL2aCt zNWRmxs^nc!A2H*5{clqfQkk-iFL5aRN1LXP1?G*`jVn`A6+|E+0@tgq-SGO4HftO1 zef{@NJvMu`^dD{7I;DmGd#4)e#(Vevz0K|sQR@HblznZ|;=i{U|K8~DfjAxfzX#%< zv+#cy2%efa&6}x-U4-7kE5P3E?4(=XYG#jw*4E!EAy6@k!>X24DB90!WwB`xtE;zT zrAA&K>ft2a!@N{Y^6%Gm?qz@#I-q=o;lE*099Og)p0oeSmmwHvDOqI?1c}~ycO`=_ zwzbspQ@IVe=nGMOzn%PGxd=qD>W&M?cvq6yqa&Gyv8Ezh!+)iqy963&4_O5&iS+s4 zPSFMaAwo4fX~w9MXr$HJ@l+zNtUaX?P!FrfX51$?iA*a$)y~9Dh3YPYFdwlS;nduAL!@7BnSxKKwQ$Qz{I;YLE$Vx{rYHHt9HCmCVEVoJGEvy5gM^smES;x#qY~D>{wGb}uTD?E2Zu8bZ6B4$+4DB5_di zd!cu}Kbp!EtPCwaRp#44Wum+!Is>emQT(sE&e4XkA|Mr&dN$@=VUzR&({m3_iIXr9 z(-R|IMoHD~1<3~#PpPD0AI`;$UMyoVxJIH8re;#T>uH-^Lz8`K88~lX+I$?RtbURh zk)3?A#LkPt;Gi7b57n+ls%C4t8b!b6dvvkqK7VY#kym-yvhx)qBVL)fJxE|1auhnHIq)iZ;OI_ejZ7m7T1!lQ~+qFm0YAo~Zj)+6>S zR^v=%ercH_A>V8s=tP}9rU7zml;;$fq03@xxL=KUTAE|+QomMm=SV1-C}(Uc2)0^M zgCPynEHe zK*!Wq^J8uy3~4UlK2z(#0lc}~vi#9kH+w#x*l z|89MJ=9T3x@_+N9J(gCIt%&x#bMx2xia$a3 z@*TmOv+V9sQZ0ep>#11~r04%pJo^6@^9J~TJ7ud* zwK#skr4ZgY;Y;x7FJA3^-}?{w*r|UEc>e=T0v9f6UIsd|J9|rB5m_&a+4{6hPsqp> zK#%(=f^Xro*LeN1zSoVLu*+Ba$kUT&#JyJywLdgeisM9uRx96Dw-#qez=U#*#?ph* zQoM~HmWYWgzaKz9vgnd?qD`xNB<;BT+%P))w328PPj*AZz2Bvo^K&<-rVCBXfh&`1na8li%43uy;;vDru|3`V%_yJekjDBwm-7GQCd9@)705F`Z8dq zkeg`O4}QJyQzV>SywBu6zIlaY-@p$$*ZMZ?C$;wb*5|lexbQ$

\n * import { animationFrameScheduler } from 'rxjs';\n *\n * const div = document.querySelector('div');\n *\n * animationFrameScheduler.schedule(function(height) {\n * div.style.height = height + \"px\";\n *\n * this.schedule(height + 1); // `this` references currently executing Action,\n * // which we reschedule with new state\n * }, 0, 0);\n *\n * // You will see a div element growing in height\n * ```\n */\n\nexport const animationFrameScheduler = new AnimationFrameScheduler(AnimationFrameAction);\n\n/**\n * @deprecated Renamed to {@link animationFrameScheduler}. Will be removed in v8.\n */\nexport const animationFrame = animationFrameScheduler;\n", "import { Observable } from '../Observable';\nimport { SchedulerLike } from '../types';\n\n/**\n * A simple Observable that emits no items to the Observer and immediately\n * emits a complete notification.\n *\n * Just emits 'complete', and nothing else.\n *\n * ![](empty.png)\n *\n * A simple Observable that only emits the complete notification. It can be used\n * for composing with other Observables, such as in a {@link mergeMap}.\n *\n * ## Examples\n *\n * Log complete notification\n *\n * ```ts\n * import { EMPTY } from 'rxjs';\n *\n * EMPTY.subscribe({\n * next: () => console.log('Next'),\n * complete: () => console.log('Complete!')\n * });\n *\n * // Outputs\n * // Complete!\n * ```\n *\n * Emit the number 7, then complete\n *\n * ```ts\n * import { EMPTY, startWith } from 'rxjs';\n *\n * const result = EMPTY.pipe(startWith(7));\n * result.subscribe(x => console.log(x));\n *\n * // Outputs\n * // 7\n * ```\n *\n * Map and flatten only odd numbers to the sequence `'a'`, `'b'`, `'c'`\n *\n * ```ts\n * import { interval, mergeMap, of, EMPTY } from 'rxjs';\n *\n * const interval$ = interval(1000);\n * const result = interval$.pipe(\n * mergeMap(x => x % 2 === 1 ? of('a', 'b', 'c') : EMPTY),\n * );\n * result.subscribe(x => console.log(x));\n *\n * // Results in the following to the console:\n * // x is equal to the count on the interval, e.g. (0, 1, 2, 3, ...)\n * // x will occur every 1000ms\n * // if x % 2 is equal to 1, print a, b, c (each on its own)\n * // if x % 2 is not equal to 1, nothing will be output\n * ```\n *\n * @see {@link Observable}\n * @see {@link NEVER}\n * @see {@link of}\n * @see {@link throwError}\n */\nexport const EMPTY = new Observable((subscriber) => subscriber.complete());\n\n/**\n * @param scheduler A {@link SchedulerLike} to use for scheduling\n * the emission of the complete notification.\n * @deprecated Replaced with the {@link EMPTY} constant or {@link scheduled} (e.g. `scheduled([], scheduler)`). Will be removed in v8.\n */\nexport function empty(scheduler?: SchedulerLike) {\n return scheduler ? emptyScheduled(scheduler) : EMPTY;\n}\n\nfunction emptyScheduled(scheduler: SchedulerLike) {\n return new Observable((subscriber) => scheduler.schedule(() => subscriber.complete()));\n}\n", "import { SchedulerLike } from '../types';\nimport { isFunction } from './isFunction';\n\nexport function isScheduler(value: any): value is SchedulerLike {\n return value && isFunction(value.schedule);\n}\n", "import { SchedulerLike } from '../types';\nimport { isFunction } from './isFunction';\nimport { isScheduler } from './isScheduler';\n\nfunction last(arr: T[]): T | undefined {\n return arr[arr.length - 1];\n}\n\nexport function popResultSelector(args: any[]): ((...args: unknown[]) => unknown) | undefined {\n return isFunction(last(args)) ? args.pop() : undefined;\n}\n\nexport function popScheduler(args: any[]): SchedulerLike | undefined {\n return isScheduler(last(args)) ? args.pop() : undefined;\n}\n\nexport function popNumber(args: any[], defaultValue: number): number {\n return typeof last(args) === 'number' ? args.pop()! : defaultValue;\n}\n", "export const isArrayLike = ((x: any): x is ArrayLike => x && typeof x.length === 'number' && typeof x !== 'function');", "import { isFunction } from \"./isFunction\";\n\n/**\n * Tests to see if the object is \"thennable\".\n * @param value the object to test\n */\nexport function isPromise(value: any): value is PromiseLike {\n return isFunction(value?.then);\n}\n", "import { InteropObservable } from '../types';\nimport { observable as Symbol_observable } from '../symbol/observable';\nimport { isFunction } from './isFunction';\n\n/** Identifies an input as being Observable (but not necessary an Rx Observable) */\nexport function isInteropObservable(input: any): input is InteropObservable {\n return isFunction(input[Symbol_observable]);\n}\n", "import { isFunction } from './isFunction';\n\nexport function isAsyncIterable(obj: any): obj is AsyncIterable {\n return Symbol.asyncIterator && isFunction(obj?.[Symbol.asyncIterator]);\n}\n", "/**\n * Creates the TypeError to throw if an invalid object is passed to `from` or `scheduled`.\n * @param input The object that was passed.\n */\nexport function createInvalidObservableTypeError(input: any) {\n // TODO: We should create error codes that can be looked up, so this can be less verbose.\n return new TypeError(\n `You provided ${\n input !== null && typeof input === 'object' ? 'an invalid object' : `'${input}'`\n } where a stream was expected. You can provide an Observable, Promise, ReadableStream, Array, AsyncIterable, or Iterable.`\n );\n}\n", "export function getSymbolIterator(): symbol {\n if (typeof Symbol !== 'function' || !Symbol.iterator) {\n return '@@iterator' as any;\n }\n\n return Symbol.iterator;\n}\n\nexport const iterator = getSymbolIterator();\n", "import { iterator as Symbol_iterator } from '../symbol/iterator';\nimport { isFunction } from './isFunction';\n\n/** Identifies an input as being an Iterable */\nexport function isIterable(input: any): input is Iterable {\n return isFunction(input?.[Symbol_iterator]);\n}\n", "import { ReadableStreamLike } from '../types';\nimport { isFunction } from './isFunction';\n\nexport async function* readableStreamLikeToAsyncGenerator(readableStream: ReadableStreamLike): AsyncGenerator {\n const reader = readableStream.getReader();\n try {\n while (true) {\n const { value, done } = await reader.read();\n if (done) {\n return;\n }\n yield value!;\n }\n } finally {\n reader.releaseLock();\n }\n}\n\nexport function isReadableStreamLike(obj: any): obj is ReadableStreamLike {\n // We don't want to use instanceof checks because they would return\n // false for instances from another Realm, like an

OKIy_gKha=xvtb+3MU+}}${X8OwkdVe)hfU6@<KJ6 z8+~i_PUUzYg3x1h=lC+Fh7{hZ$vy5WF@|`5VPn{vf}O7@bN$1wgRV82Z})tMJXJK6 zE1|r#Nt7>g;xLY79r%N(+mTcv#R1}&An8w`^&=MEZxbFdZ0Kq|TMC(8u_2nR*N~Lt z6M5BQ%Pe)2VXv#_Dxae{S8acF;F*~?{H+?BP2uUcDxt&lcF~D84C)IH5z&mW|#^6IyrzLWh?Ngm(ZE;hwCta*lwzE-061AAv8z!H2Z%vt!=Jazp7 zQNu0H`{h1NyWqE!EW48cqf7Q6bPVA@O-QIo3=EU8c@Z!Uk56zGpMuB6X{BA90u&)r za802QUKUS)Rb~lVb#)s$^Ps$h=h0VYhpyl7xVRa2{tV4mL`OR&x+4g9b8Y*GbXo#Z zKHiy2ICVGN%=2h=A(YrC`-SLOBMl^Lno)`jp?EQ8YTfHYvZ@X#-=rl zVDokI;b5W*Eag44v+(zX6IAi|Y?SIA&t?R3htTe`#;o0Uat=OL7vQ;s-=5|$e;!~E z%94I5;CWYb*F@xkkPf>-EomIi=dQ-%tt3AVZ@)Y1Klk(70P(Kb9I1mCMMy9X<(`87 z;N?}8-+X<&Ht}*R30p*M`Q-Vs2{hR01se!7MS70k?!u! zeQy8ul=)*?bo&N!T+6}=wT#JFIgdmoG7$S|RfaCN0=Dw%hirc^rGa8@i&K=@5{-71 z+*|y4*|#GbbgsCr-T3lK#e%-`(6uS#C&T-|2dnrcSV|a9h+j*ihSX|g>TLz|xebs8 zh(gX^Y0oOl%63kTs?fdF6P|Xtb=se7&MA1r*|KO+o~r^Cal{NF`ig_YonQCs<-!Wu z_jiyyYB&Gf#b$pyy-h(XEQD88wBSl~Qu1^JfdLC~$;@J$g1?UKjxyGJCN%{Qqna;* z{7W*Av1T%((e0JNb-dUleA6LXs0|ts#JA6BApbt&&ZhKwUFKCxC4^e1h$S9^M_Qif z1=~FL<8i)lKy|^^yw8mB$cfI-wpj%?eBYX+;aN-it}_n5E>Ne&2Q@TgkP{+EVj*@> z@4E-t>ksW}`?>CP4F?CxE%YyT!*rnN4elx%dxpm1>!2;9g9U;Z!Y92ROX1 z0W`g>0S!reU^KOtCgw!%k@By%D8t3Ds?H7>K9wKpGG*w{m8%4P`kJAzA6^-;_je( zT+G!=`xn^Nn^m5YIW8e^Xwm+kHWq~oL>=@jk&+?(399Xbd9zm=2>J+UPqL-MPkKz! zbL8D682F2OEQ#>L7KV%sG*S|Nyp*!q1_Xi0I0i5#NJ+%Y7d?5pP-R*mfm`|U41DMF zGX8_hfDW+te)-&=W@PJrztcFO1@nl$-+B|U+q2)HXP{jenSm@1%-%g85AH>b3QeB8^tvx0R$ic-IAj`9-# z`XcP3cYnwT*bYK z_)XHmq~5rVD9S4sUgOQEP zv`6;!H@8z?9b-#m&Z!x?23-BoYSZ|1aWsV07I66`mlq1XD8RFV01GqWm8f*us^~$n z7jYjyd0BOh!>{7WOX%!8g8_&+n2l2%U5u^no1d&@T*38R)Uzr#j)AIoipC1*tPa)BO_sPIJ65oV+SBl@8 zb*MZ*B5W;LpIb980R(|beviC*ZQ__@6cuL#lZ)?9Dd_$?+Uy4?4QV~D}V2jLe} z65J(MS62s00Q~R$I2Tk6j>QMFZnU_iF~2qQxn&?M3hQ$NYWI@3UhA72xJ-Pklq&mk zCpZ{=6OEi0(XIFMs79tp1GytX4ui^gPExB87R`8;|m#AB!tv%mY&nDC&Y+f-xoRAD{8vy6rwOczbt$ zTqP!d5o5rWc{R0GCCGQtr8X6DAeJ$I0q5th?H1lj!f!ah^I>8345s(ZfLRs7;Inub zp-aw_Z*3ii!+HI*OzGL@5JK2D`qGuBu&9Sz_WR zFc1=?zmODl)UL4%)75}GhNA~1`~1ws=32bs(m#`Q5{(17b0un}{LX((;1N7hkm4#x zWbZOhyv!5#Aj9;s5{3D!ADP0x+M~$$kM`gO7$JG}r9yAN0yCWh_9sny`Rrl5q=Zi# zDn`O-ro=F_Xs?(1!zXdvZ+(nLlRgMa{4&S0xPOn-W$jOZ!STgF7$pPS%b&zt>`c<# zkPn||W^HkzRgH86B6fC}{=ywh^BVKd97dCBb2GPAT=7X{ci=uo&;1&d4XiPPT@|O; zJ;9_f;-7HfDYoto165B@mX@DM>$7CoADKJJ5@BV&@OBQ80f=RWbptUa8*Yn`+)+8_*0`A-> zpTF*oX(|u6w@RvC5GxD&N~FB+0hG~@3q97}M9nxJ4w72dh34)s71xyZde)e3W)N|X)Rs1>`;%}$$lJ6Ea6U`k7y@ZJ0NP%m!XvZW8 zB{?g%9M!oy2p>p_>7*KL>OeW+4~ldSmjIpWWeDNIP^MIY$r{w72ZFTCez`ez_~d(tHM=(YTL6qo=&B+dfIxQ+FK;Jq(C!s$R{J zpJ{;jqal(|RFFUJF?UCimAua? zJ&|fEU0|Ixhwn10y+e9(o>eso#zA>7ER5UR*b(!WCDm@)nIPM2K5Pskx>4T><<(Tu zXE|};De4&MdFMFkQ@Ud*i_LN&hui&$obmF}@8rqSxJPZIMVxbWM;{L2~et zUcEqLf1l-v-M~j5QVn!wop28XdfBdoyrfT`x%orAPSU;j% z-)VRYY?{B;2>wM%(pJWahO!D6zcw`$9EZmZ+4Q}1^9BY|V?SN%{PJ1 z7W~_tj-AAgLpRT+=vJ=9XPxA-`cL7%Lzd&|{jYhJLX75|I~P2aQjx6n6za(wH^;$A zW+?Hr+5B!q^%t>EFsdFLEf_WhZgJs+fOO-Vr{<8}b7^bR<=F^g&(rVBP14Pu zNI!2P;i=~o%6shFZXGhUfh_Tp2nJ52I`Y zz7W}l$f~?2pk1T`^7Yye&wRTS;}a$y#E^DxSweynes@28^DIo-aVfY7L(3yoGrX$K z)rpDwJ(4=VkbdHL#VAx72#@PeQoY-RcuxvlV^Nle>k$X&NItq7JpmtEh2<$4pihf{ zivbc7sayE{P3Qh~DA_OM@kAggL!xiTn@STJl7DvxO49na#4G@lbe{?Oxqy$kL<~S~ z;jdIvT6t#jV#m<4H=B83^Y44yXI?rspVSW&Z&15fWz#ie63+6@kq831Ep2anE0H4z z=$Hu|VK|8Omi*BthOg~qGD3IhrUQv`_ep72R}jnONL1>1#en!)P@UqDG(4KJ0@9}D zNLS4s$;qry$zx$lF)|-k`duvnBS#U}{Vq}mY0^84z<}@nRPywFFQeNLZ1*}LX2<9u z6T#ic*eONwK^eC#kr!F;{R_ijVkLBU33o_-B}I@$pXP0ug?lNx&xMX=dn<^oY?k_& zkg_I;#fJw&uZGMes&z?8^n*~TkbQO6G7D*>IMBlez}I!5weNZgcTOTh)`gJBu!E;B z+k+08e_c_Tt0lgP+F-4a9&?x)>;ZMS!D;L8bwq^6H&WryW?Ufb8nP{5+U?uPdWQT` zcNcx&nad%OzBB*>_f!Vt4Y0ZlsNiSX+H~N4!4`m|ML-{-CU3F1>{FwPMA5e;n6r_2 z5AHVDwTGiM2RMdOazsrVneoYn6xdHW`b;>F>hJG`%NpYwczOAUl~5_^fb?6se^8)C zUD9ko;KV|~wXtLUN`Ln2ypuCD%HhPj{z<`1frfj9+hdLnUwW^xZ6si#o>>t?sKuC+ zHC)RcOZ(OVXH_?$_wEcZNP{nu6j4rr;*(-e02ep`-*lU#R+ zmF8#Jt=<>W;XhDuAzHL~xg7*yN{|{K{T~dc2?*j-#xjmIFgx{syu^VH+;~gI`NhoP zSu4Ec;3yaIu6K77!HlvR-3YO4ta-cnXS5an!6?snA7ct{5|_Wb-92$5I{ zbF}TG-rM5N;msFgSS=OPGqwQ;lVR6OV~9t}ml-2t;XhMGBCi5TFo`jYiJ9OMyK8_m z22%JNNQ+K4slYpuv`L^vSYsS0bcwbC#3dTS?A+{X@4N+9u!LfAEX)$Vi`S{JGIw=q zdCOfpWVZUO%oVh_3P><)Lwn=zr_X6Vqx}V^*7X#yAKnG zPq4w9xd*cb?mdS1{J=MxO}y);he2csQ(8?uaSeCq=M98z&{aqdfP3kE3%xe>Bg*mY(G_C7qzdhx&}O*@cYqhotyXw@E{UjUB9@|<4nD?+cP%QL$pL`&&EdL zUATub0~bW=pf3zF(~5>KNc5J2lS$D_eYH+03Jk>9^4ekxpjagB<2jJxvjYkJh!LPI zWp%e3isC&8YFXRK4Dupibt~are`H6Ho3{c;uwYz-vQbBM9`sn(c(ZxJDGClFJB+aR z-uUyB=%>KdgNZ1yd2`|oV4Wv7pvgj4uD@Ejzcj>RX~I5z1RaCv+p*M<_(H-joo=&_ zAX^NZs4>=S>`GS~z=x|=&ZzBCV56{m(s1lz+|w#|jRPv!KtmLBw#eH2i-w9RewdXiEV(dl-5an^IvE@1Wd z1J9@5$pB-dyj=gr~FRv4Z=@!>FOzV1jGT;*h+1 zKL?KCW!t-(g7W$D)NWLoE(E#en2`;LySDIm?ucAGe*wXQ_GRgd(i(^{TN$)`2AWGg zNP#=-L*vU>vvl{A0oN-&osV@V;fL1a04|Mbpa7Z|9IT-tvX7A;Ab6#m-55w-T-(5| zA}PRx!I8>i;88&SxI-8;&9hFAyWD*PKv)J2U2Jrp{fI>OOG_KNGri=I&wE~lJ{R}{ zuS+=DtHBrkq?~`@hgJLbL=6@5nl zO%t>9ox37@98^@AK>O>czSJc@BJcfo^1@=}2l$E-4M$2k0mqGZ$XhTQPkXzQU_WRLqrWqeSGOrpy#R5p%3HLm|ReeO3<>*JDS2yw` zBjx0}V;7-e3pW4Sr>SjUizNFf^^K<#P_Uj?h!5PTYRsrGL4t3n4R;H3mfy58BXtn-53ck3N-82{VO3xAKQ!?&m-6MFy8IslG9otpb7vg7C+| zBGBFKfm&d~rNr(gZ+jMpZUB*<^F-*kZoc^;o&cJ9loH_r?Xos6Ez!WO;H04wimFZc zCR|_Sv2Ecu(0~Z90jdww%=W^F=v#L!;b9piNmKCeZm0g2q*%~E78-rjPBe*`BXN9S zH2ZPx)Ekn{pr$VMa2QXu$|gQxb;?5xI_7xQ5wRwoRtZ>}k8 z#!*!Xc3+*ts}@l0@0hip3|#5|r26@JS^F`{dAp|y#kTBHcY0Ea#;gZge$jpEcyxc{ zea2>YuaAG>48~+Dg*Ns_4vUEdqNnin-B}G~y`DQn%1%V1)fvK&B1-ygF!@q{n%Kwf zH17;F5I1!#5m}>ueOEooK(Ob8er>%Ru%hDcws^T*9%Mg8SvR)0?Aokagc>E=HX?_P zWxNV5UsNX^Msm(701*irG$Z3H^h=As-rvM-6Vr0 zKc-qhlQG5$z`V5rh&kIR$vNARBWFj)J7lj*Z&t)S-dXe0WrJ?GJ|vs_`HJH*dAhmy zgrVf^>RAz9pgodm1(s$jsr(aOfLF}O;lS?9Vv%vTps+*VTZhFxY_J(w3-`=c!Z*bM z2nt_F{HCe$%^3J?Avc~feZwj+%{4P>9fOiUyg5F`LL^t~s<}UHhea`@H_imO;C2gV z9YWXWUOuMSz~4d<(FFzN_Ia|n?{Cjs9M#QSv(EUn@XMd^7m}_69roTGcq~2U(@~CbOr!XI(gxLyqo_U@04&kbDFo-2{6B zFQU*2!m|qv+i$Y>GsprS&xdiq!n8|fN~?wRfLaM8qD z1MTyPwsSY?j#6{4u=Iv~7~T4>f3>RNy{`g%9bOE9M zCSf%DSoU2dDZi=r|wl6t>)JZvX7kRORCfQJfTwNgql!f4#ds=EX6@ z+w;ZMZc9W_PB<~gXyD&(_pw`{!hcEfs&`i4^rD;69LuXS2qQNeaJ~L-lH82NqyDYi zF3s^P|20y*L*>gJ|J6l1@x~wjCaKx%1LwbWv2yuF_P55CG?D6|M&)9R@lW|tGFIv=L{ zo_iAtdL(H#p=LCj&>8mcBE@o_xhuRiA81#cX(cgB_z@kI(RHDxPwOA%DR{Z~Q|(^q z|1P5vBrul9HzajJ(D46|F>prG^zMPd>*$A$uQRHfpHdcG#_)bqaUxq+(K+#piBC%T z@=~RJU9TF+-$_O?On*o+IlQcxu2dim1XQ0j*bkfSVz@paEDhSBl54CxU*+zyhg72S z3^t+ZEE+NP4Evh)nDyIrr@#cOA2ICSJn5}%fL;<_dp{!`j1Mm}&EFON9*4i@!(Yz@)duox!V~UbN-Aqek@1Z3&otwxX+=iF4Al>c5%Kn@ zSp6Afw? zg3h^s>f=*RD;gilKRomC7)L|X^L-*RA_`~W>sCbee+j~roQdQCIH zqo5%u5H+XERS6oI9(KhZuvXTRQr~&L68Xhd6c#QIns2Px`Ad3>mM)%@frpJ>>n$_& zYzvhe1AV0zq#{0rJx|>2#>vS%CRwAT0@s?V5jeuc``Ih z_sG=t+pm?Br;XPo)BFxFwzaHHiBUs3wD><1@*~Iw^=IjZH(x>h&s^)FZN91oB)Y!XjpNh@mvWCS{?+B4w;r z&Bw0CU3%H9aA<~o%EAJf_0Ip$hOXu*U&M{7!Lu(+(9DUkcO|>|Z)Th3KM<1*nI6TQ zEn4^yCr!PASBn6;;~g9JKB`;#`h3m|UjaY>nc^DgsIWh};15c-Vaf*EJEKEF|1T2*=Q^1L^*;46q4#%D zxRAzp<@4dM?vJi5hM@O?uRrq(l7JLS1lmsIMt+H2iEZ9ngx#ikyTK85%0X>((Ks0z z_;Z#xBCn1icE9(k`#MP+A$kQ?&(rZnC}rG&rMc2`JOtrUAJD{KsL6fVMFKRSU%)*WV{~F$P>B-cdUPwaA_|@H%HqOUi62()UyY3BG319jh zo)ZQ#A1kw;p4DwXe$QzmQLaCRaI03cu&fe|A8H3ZnH?nW@h7K6n#xsrr>ff~N}pv4 z{@C%UMZ`Z0B~>@fH6m`iomnU~b#i&DabtePidaiBg>TgLO*p4{MgK`_I$uljBMyL+ zC5_VZAv1|aTy!d;A53|`M_4GfB{<&d6IMY0MBA@gj z_q(grhPH>@f6?iIFpS50+PdDM zTBY*LWJvLNS#myz|51J-VAts#gOk@kcg{gyw-T!N6&%dC|HduOlTu#$Nq7S6WufG< zWkTD~cj2d+?~i_=2-=p+DO%1Uk zxQw08)jRGH_cKgy-A*`#5}9=6(w?UYS!L3WT>8$8Mx;ic+jrnOp-Kjp%2Iz_(b)#+ z!9kSOD~!B>+BJTVMkKvvOY9(}`0L+uju*i4X7K!mkHcS+9nG3C+2s2crUfp=_y{zRWx*{S&5A0Z5s#WGQF6K?%9Tlm@*?YVwg4IPCdx>4?yskO8@kc< zcr+<6`IPB%1dJlDz{=F-Nfd%jfZ@3i=2Ig7)SW{{F~eJ~SUMZDP`>>L3~RA3-AfZ? z*+yA~f;h>=o2P(a1{gi12PNfU5l0m3_re;enDaTH)5tg6vR@c(EkAwq{OZnwnA0~s zA~oft$;C)>?a#es%2BQ_3XfUFV6|^w4pH4p5C8c6XBv$#QB~${oEoVQiiM`?8b-!c zGMxr)EuE>3yKWNTR!Z!6!zoTfkUYKCg+x2<7{{D#fz>KU*6K=nHwc54ImiaXho2kQ zDvH*{Ey}Fs7?MvLft;co`!W4eNk(7s9l?myBJzdc#K9j0H6T@^q)RLzz6wyf#(Wo> zD7VAa_&KN8iJQ_e#cP!@IO#voxFwawKDa>If&OdzuUT<8bfz`YY=Y{lzNV4cI^Doo zg6{Htka2B1oikrvt!miQ5!|J0L1f%1tDV@RuYNh{#^YS0T#&OP87LTBYaVvphW@a< z^A{lI?qQo+1AH8`0zUk@s96+m==jL-6ju{-j1|B(50N;OHN(MnYk~9fn$!gM0lFiY z4!!lA2`!&XK9Tldz}CNrvwsL2FC)skM-HY7l>Pqw#=I&+!@Kd4FFcO#67W^|>g9sQ zvswYyf(P20wGT zKBAB;g+Lb(T!_NK*sOolCHn5S6WIvQ8ZK-~gLV5Q28fhX9>9mq-oe}2oqOvaw!R0R-okBG1^QcS<^n3`Wg9=uCi%>C8ox<%q8MsqUOPpTvOwRu>*ddq zhO3^g%uf`Ii|J)%AHHiD2Y-swtHJuPZO)?PTh4fYt&HnC+zayQQ&j5oDx@#OMyxAU zyV{8L^f#!kOj18*0~zJhE8)iL8MfAqU;h7)c>TXi)XVMp@Zj09Y1WP2AjObv8vsZ6 z1C?B|NVuQyiLu~Ml%^ft@_gj*fD!!z0fij(=kg{4O+R_$qMgV(No7M6VHAgUS^**t z%RN#v_j9^8Kp<4>;x^Rk|7q_$qng^WNK*_DdPfjc1SyKh z6(YSDx^MxJ2uSZmkS2yEV2~mmkzN!;5MoJ4GCS&fuj`$eHEZ7Wu9+`4=Tpwv*{7_N zefIPG{{QE680o-)T_J8UicnP$#V%s6o4~1=b!tmtj+qC&LdklZRjMY3#a>f2wYzGC zeb`E}vW&PlU;NdyQElK9wd!S`K0LKH5K6fi<`wlGw~$`FFz$4y3y%=74T-^!5ZUGn z>qVxUJf}368So03i%Z_c8mPu3i{-tqUjrIK{o>#m^)V&HEjx=q?3H_?Myk%0t1xuk z##T*TGtSqKsAp8Nj3f;&VP(lEI%&x&X=zd4T9Ud>Tv}g@{>)TNDhB~BT3`;QLTHjc z+Jlx*wc~7Ds2Y!VZ`{f7AHVr!(q^`m!K9ecPPHtW=EKL2@M6-CsCyT;0f9>Bt2#J79>!N0fR%$yCW%{kTuK8T7OHY#VLNLSiaAVMj zwl5jSaYZELvtUwwMI>=zTDrD_`3vCu>V5d(?AX>!5k@A{&w}|f%2EJKpbjiV61n?8 z9u2q}t3EdD69${Zk{3|hcpaXa0rOj|m!SAn&W4KLa(Vb58Ak+Vcl^l*RDL{acD!E_ zUwm)vLuS{zCXd;I1p6~?X+C)p_qnl#+z-_uU8{QbYnx-?Si;9|t39!>`QK@V5EsYI@e_KuMqn>gs8OT@EyMrkhKQ}R>A<@O>4tJrKU0aU?$&$0_ zR?*_&E4>EyAL)06w)_Sk>Iqcbb0(>)6zgt+|H<34 z=^c@JUb>byFdN{--?pc>@u?H&Hkf+ks}11B<@rFVSW zhbQB05z~pi>ODc&My5CjSMI%!I!{l3n>ac>z`8moEDv!eURnHF^WgNRrlr*;_>(>* zZ+(<)B`x64ANX7N=&Ll{Zq0h=QMHm{?j?$*1QIbDY$XDY$Q}H-5$DFlPZVsG&eJl5 z*+|T zggzJ5f!IVY{2dR~K;3TEyKuC-@zEm_P+(1_%rz9d-dfJGcg67VH&9O@BTF#}lW&ko zCh6K5PBMp}ltD#Y+b6Tpkb`Ct=^p(~w}xya?*ceJh6tTJgdytpyB^d3@SHOA_>os| z(TaT!l`tD8Kc=;kvaPIxH5YKuDou}o(StAM)^S$Ba=J{~51@~W^?zm6iHEt{e*w-- zBw^|c!!Xl-5O_ALwT+LQ{2Why67`gQLAm)~zJKz$x!V4V48Y*if06}cl7H>6kNnR& z@F4vv8xSK$Gb{fpBd~w??f>uqVnxu*vD9&-GKiYFQdliPi zPO1)ExRQ8r@>}zwxczVll-crSiF3-g=*2U76kTM!vTcRX*-2!+=Vi^3oDqx9P6aUp zw%Se(TFWv?HXbVz{%(-;FXJ8d1H;?72m-vG@dM`=qPd@_FS$qx-`tJ}9Se%1LrE7W zECN%yzI@&a@L`-}2a`9cdW1Rrozxe>KYQGe6p*caq#8#Fqz7Wsr3W!Ur|!rp=Go?3 zo_IpgC1hFF;cx}jx;0i@Qy+45Kaj zBU&H<3wZvZe^DxFvwgEYnQ^ndn6nH5w9bl(5iFv#n-PPxdYPp;9VE}koZ>yKfe0J> zgDD^+I%T%-4*xN;)y3t$a9X+Yz&b=#7S7~a<0?P^Y9IQVyizpumJ_XFjB!4=_Ta_cQU`|L+8k$eyn~W^8_>H^S;?I`1Xj~pFW>xUuDo&5d>dDF4OW3YfB&xz z<;f9@t4^2hUAg3*cjHeio$(E7E=!iOJ1;HIQSjHku$sAH->jxC+Gb`-{6(zYNeo=r z_Wr}K`?0$3Dl4HvT#1e-L)l$;u`|h{A;Z%Bn9;498V5=O_b#3f%S4_zxLZ}*T`wnZ z+FZ85V9=eL&4-`o{`oU~@PR#sc#()msq5ad;~&uA~T>p=H@-J1o{$EHwb zcC+(1+xOot-Zz{WgF*6s~+Wv6XrE!{x{StHL}6lL2~;@%iEi^i?}_lJWpT5*p_%?>#wPO~D@ zBPIgPyna^n%ER|ZD+`anW?jiQWNxGpPnGFib|F_B*AJBJd>T4Hx|bd5Mjp$dQc8lRX|V=URk6owQ4f=k9fpKYkaH9;Ogu}ZvL+A*Er-7hNvN-B zrKDX?9RW3iw%Ydza}Px4$n*lOrzUy5IvOCbKq41k3DYJB&_>?Oq>*WiGnqW$fZjUC9XTZZUR~5VH%)a1a*|msbOZdEfhKWN3pW$4!8(lTu{;~(jBMcL$At!7%x*2+b(fpTrdb!&??lm zG+2&~d!M`NMsee8R*z<=xTfdGST=Qnoc48h01AtRw%XqcKX!0i(((^27r6MnyG#rE zUA&Vd3{)`!s{6Te&d+m~Pi(fY7XZ62#6~z-P^vUK3n4U*A|%SmsmPy$>zVn!0O=4o z=uyy9+*WN}e6xMVaZ${I4Q7`@pG`C;PP88@tH_C!c_x2n|9u{vJN3~=rJ1T|jnsO# z6Stb7Ije^t>@9je!Jh;f7E5v&>ieYO8T}8jf_zzrcL%=?ol6sgwZzxiA3dUb0jmi_ zFJVPdMct*9KzMa-EMel0r^L+86^rc6<3PnDb#^%mrQnqpt)!$o_tXTWwtcyOrwJMPLx+i{HqxxZ z`U)>|x5xt@e{e#B`*wdAk2{pz*TJ5(=*Pbvuqb|6W>HY1@*{7{`ujP`rs5*5xbIxm zDlUuw-(apNqvS1K5*x>&%8C#5vS=OUlJ<(P>pk}7=ywKi-9gF5uELBJ)&`jn8W*fW z-X=<0U^eE)kJwr*kO0KUAzi?ETb=L(_`Em4a zqL>utCo~|tF!E0ri=#85T+U>sw%81|vt7g&|C{PI|1u>WTp}%@nTPe29nS_&HJ8~K z4;U$1VC<8;a^79FMDXn^aQSOD*c5^dUfZwDu?MHzGb`N{pJ3n!7`NMlOU+qhqqLgF ppzw!R`rkkJ`zrlUw?^bXa@thuMCcsX_HW>0c=m#BGc)JR9QMau)O=X#!i9;T z37kk-upj^(fsR36MHs_+1RCI)NNu9}lD0S{B^g8PN?Ww(5|~L#Ng*g{WsqleV}|#l zz8@ri&cTzw_h33bHI+12+kK6WN$h#n5cD8OQt`5kw6p~9H3()bUQ8OS4Q4HTQ=1Ol z_JAocz`fLbT2^{`8n~UAo=#AUOf=SOq4pYkt;XbC&f#7lb$*7=$na!mWCQ`dBQsO0 zLFBSPj*N?#u5&pf2t4XjEGH|=pPQ8xh7tpx;US5Cx_Ju;!O`ya-yF`)b%TEt5>eP1ZX~}sjjA%FJF?h7cX8=b!DZl<6%Cv z*G0uvvU+vmnpLZ2paivG-(cd*y3$hCIcsZcYOGh{$&)A6*XX&kXZd3G8m)G$Zz-LV z^GF3VAW^Mdv!)4OM8EgqRiz~*Cji;uzl2uC9^=8I84vNp;ltJ|q-*uQwGp2ma6cY7 z;`%`!9UXO@fr&Ebapfs34OmS9^u6$)bJxrucutf>`dKPKT%%*d3XlFVKunp9 zasduxjrjs>f8V=D|J=XNZp;_Zy^WgQ$9WDjgY=z@stwiEBm9u5*|34&1Na8BMjjgf3+SHcr`5~>oz1Y?SW^=K z^bTyO6>Gar#P_W2gEMwq)ot3; zREHn~U&Dp0l6YT0&k-wLwYjb?5zGK`W6S2v+K>AM(95m2C20L|3m~rN8dprPr@t)5lsk9Hu*W z?pS990s;Ez=+Rj{x7p``4>+c0G5^pYnB1^!TL=(?HLHZ+HicG{~4F1d^5Awl_2!1jICM-!9eoLhbbT^;yHcefyTAaqRcY zmuctDopPT!%k+}x%lZRKnzykr2}}XfG_ne?nRQO~?%hkzo;@RN{P6o`&mMUWBYMTe z6i8ChtjX&gXl`nvrU>jah)2iNM%JdjqoaeaU%yVn!^70x-flljp6Q5tK}5}&X8&&G zX3fpb3E(!rH=zVI_9Gjl45w@{(ITqngWFe7@9{mX;tO25Z_8 zQHEpI+FkTU#4xu>RkN>b3Tnc3UpWzPXWm#o55GKF09j^Mh~)K7{QqbO_~(@CVq! zS<8954|P8mXN2MRs86xZ&Q4EfM@JB94b=(YGuk)s&^jiSF=t3*oNK3`rD{H`yQ?d; ztE=laAUoZx5?RC8*WKOj`%LXEkgDd>&^Q4M^z`%u0rg-It=hLCVsq!Z%^6eB-OvOT zFZ28TN&cRmgU}Elrnk43)!>Z1FCPL2K$7}gwzIc48NX}#!A1BpJP?#v5wkNprhV** z?Cpalt1oH&{r!o3eSKc&ap)iz2BTn_VV`4>9M^b3;(YY}4>#ML6{~(4mH+?%07*qo IM6N<$f(jP3KmY&$ literal 0 HcmV?d00001 diff --git a/assets/javascripts/bundle.83f73b43.min.js b/assets/javascripts/bundle.83f73b43.min.js new file mode 100644 index 00000000..43d8b70f --- /dev/null +++ b/assets/javascripts/bundle.83f73b43.min.js @@ -0,0 +1,16 @@ +"use strict";(()=>{var Wi=Object.create;var gr=Object.defineProperty;var Di=Object.getOwnPropertyDescriptor;var Vi=Object.getOwnPropertyNames,Vt=Object.getOwnPropertySymbols,Ni=Object.getPrototypeOf,yr=Object.prototype.hasOwnProperty,ao=Object.prototype.propertyIsEnumerable;var io=(e,t,r)=>t in e?gr(e,t,{enumerable:!0,configurable:!0,writable:!0,value:r}):e[t]=r,$=(e,t)=>{for(var r in t||(t={}))yr.call(t,r)&&io(e,r,t[r]);if(Vt)for(var r of Vt(t))ao.call(t,r)&&io(e,r,t[r]);return e};var so=(e,t)=>{var r={};for(var o in e)yr.call(e,o)&&t.indexOf(o)<0&&(r[o]=e[o]);if(e!=null&&Vt)for(var o of Vt(e))t.indexOf(o)<0&&ao.call(e,o)&&(r[o]=e[o]);return r};var xr=(e,t)=>()=>(t||e((t={exports:{}}).exports,t),t.exports);var zi=(e,t,r,o)=>{if(t&&typeof t=="object"||typeof t=="function")for(let n of Vi(t))!yr.call(e,n)&&n!==r&&gr(e,n,{get:()=>t[n],enumerable:!(o=Di(t,n))||o.enumerable});return e};var Mt=(e,t,r)=>(r=e!=null?Wi(Ni(e)):{},zi(t||!e||!e.__esModule?gr(r,"default",{value:e,enumerable:!0}):r,e));var co=(e,t,r)=>new Promise((o,n)=>{var i=p=>{try{s(r.next(p))}catch(c){n(c)}},a=p=>{try{s(r.throw(p))}catch(c){n(c)}},s=p=>p.done?o(p.value):Promise.resolve(p.value).then(i,a);s((r=r.apply(e,t)).next())});var lo=xr((Er,po)=>{(function(e,t){typeof Er=="object"&&typeof po!="undefined"?t():typeof define=="function"&&define.amd?define(t):t()})(Er,function(){"use strict";function e(r){var o=!0,n=!1,i=null,a={text:!0,search:!0,url:!0,tel:!0,email:!0,password:!0,number:!0,date:!0,month:!0,week:!0,time:!0,datetime:!0,"datetime-local":!0};function s(k){return!!(k&&k!==document&&k.nodeName!=="HTML"&&k.nodeName!=="BODY"&&"classList"in k&&"contains"in k.classList)}function p(k){var ft=k.type,qe=k.tagName;return!!(qe==="INPUT"&&a[ft]&&!k.readOnly||qe==="TEXTAREA"&&!k.readOnly||k.isContentEditable)}function c(k){k.classList.contains("focus-visible")||(k.classList.add("focus-visible"),k.setAttribute("data-focus-visible-added",""))}function l(k){k.hasAttribute("data-focus-visible-added")&&(k.classList.remove("focus-visible"),k.removeAttribute("data-focus-visible-added"))}function f(k){k.metaKey||k.altKey||k.ctrlKey||(s(r.activeElement)&&c(r.activeElement),o=!0)}function u(k){o=!1}function d(k){s(k.target)&&(o||p(k.target))&&c(k.target)}function y(k){s(k.target)&&(k.target.classList.contains("focus-visible")||k.target.hasAttribute("data-focus-visible-added"))&&(n=!0,window.clearTimeout(i),i=window.setTimeout(function(){n=!1},100),l(k.target))}function L(k){document.visibilityState==="hidden"&&(n&&(o=!0),X())}function X(){document.addEventListener("mousemove",J),document.addEventListener("mousedown",J),document.addEventListener("mouseup",J),document.addEventListener("pointermove",J),document.addEventListener("pointerdown",J),document.addEventListener("pointerup",J),document.addEventListener("touchmove",J),document.addEventListener("touchstart",J),document.addEventListener("touchend",J)}function te(){document.removeEventListener("mousemove",J),document.removeEventListener("mousedown",J),document.removeEventListener("mouseup",J),document.removeEventListener("pointermove",J),document.removeEventListener("pointerdown",J),document.removeEventListener("pointerup",J),document.removeEventListener("touchmove",J),document.removeEventListener("touchstart",J),document.removeEventListener("touchend",J)}function J(k){k.target.nodeName&&k.target.nodeName.toLowerCase()==="html"||(o=!1,te())}document.addEventListener("keydown",f,!0),document.addEventListener("mousedown",u,!0),document.addEventListener("pointerdown",u,!0),document.addEventListener("touchstart",u,!0),document.addEventListener("visibilitychange",L,!0),X(),r.addEventListener("focus",d,!0),r.addEventListener("blur",y,!0),r.nodeType===Node.DOCUMENT_FRAGMENT_NODE&&r.host?r.host.setAttribute("data-js-focus-visible",""):r.nodeType===Node.DOCUMENT_NODE&&(document.documentElement.classList.add("js-focus-visible"),document.documentElement.setAttribute("data-js-focus-visible",""))}if(typeof window!="undefined"&&typeof document!="undefined"){window.applyFocusVisiblePolyfill=e;var t;try{t=new CustomEvent("focus-visible-polyfill-ready")}catch(r){t=document.createEvent("CustomEvent"),t.initCustomEvent("focus-visible-polyfill-ready",!1,!1,{})}window.dispatchEvent(t)}typeof document!="undefined"&&e(document)})});var qr=xr((hy,On)=>{"use strict";/*! + * escape-html + * Copyright(c) 2012-2013 TJ Holowaychuk + * Copyright(c) 2015 Andreas Lubbe + * Copyright(c) 2015 Tiancheng "Timothy" Gu + * MIT Licensed + */var $a=/["'&<>]/;On.exports=Pa;function Pa(e){var t=""+e,r=$a.exec(t);if(!r)return t;var o,n="",i=0,a=0;for(i=r.index;i{/*! + * clipboard.js v2.0.11 + * https://clipboardjs.com/ + * + * Licensed MIT © Zeno Rocha + */(function(t,r){typeof It=="object"&&typeof Yr=="object"?Yr.exports=r():typeof define=="function"&&define.amd?define([],r):typeof It=="object"?It.ClipboardJS=r():t.ClipboardJS=r()})(It,function(){return function(){var e={686:function(o,n,i){"use strict";i.d(n,{default:function(){return Ui}});var a=i(279),s=i.n(a),p=i(370),c=i.n(p),l=i(817),f=i.n(l);function u(V){try{return document.execCommand(V)}catch(A){return!1}}var d=function(A){var M=f()(A);return u("cut"),M},y=d;function L(V){var A=document.documentElement.getAttribute("dir")==="rtl",M=document.createElement("textarea");M.style.fontSize="12pt",M.style.border="0",M.style.padding="0",M.style.margin="0",M.style.position="absolute",M.style[A?"right":"left"]="-9999px";var F=window.pageYOffset||document.documentElement.scrollTop;return M.style.top="".concat(F,"px"),M.setAttribute("readonly",""),M.value=V,M}var X=function(A,M){var F=L(A);M.container.appendChild(F);var D=f()(F);return u("copy"),F.remove(),D},te=function(A){var M=arguments.length>1&&arguments[1]!==void 0?arguments[1]:{container:document.body},F="";return typeof A=="string"?F=X(A,M):A instanceof HTMLInputElement&&!["text","search","url","tel","password"].includes(A==null?void 0:A.type)?F=X(A.value,M):(F=f()(A),u("copy")),F},J=te;function k(V){"@babel/helpers - typeof";return typeof Symbol=="function"&&typeof Symbol.iterator=="symbol"?k=function(M){return typeof M}:k=function(M){return M&&typeof Symbol=="function"&&M.constructor===Symbol&&M!==Symbol.prototype?"symbol":typeof M},k(V)}var ft=function(){var A=arguments.length>0&&arguments[0]!==void 0?arguments[0]:{},M=A.action,F=M===void 0?"copy":M,D=A.container,Y=A.target,$e=A.text;if(F!=="copy"&&F!=="cut")throw new Error('Invalid "action" value, use either "copy" or "cut"');if(Y!==void 0)if(Y&&k(Y)==="object"&&Y.nodeType===1){if(F==="copy"&&Y.hasAttribute("disabled"))throw new Error('Invalid "target" attribute. Please use "readonly" instead of "disabled" attribute');if(F==="cut"&&(Y.hasAttribute("readonly")||Y.hasAttribute("disabled")))throw new Error(`Invalid "target" attribute. You can't cut text from elements with "readonly" or "disabled" attributes`)}else throw new Error('Invalid "target" value, use a valid Element');if($e)return J($e,{container:D});if(Y)return F==="cut"?y(Y):J(Y,{container:D})},qe=ft;function Fe(V){"@babel/helpers - typeof";return typeof Symbol=="function"&&typeof Symbol.iterator=="symbol"?Fe=function(M){return typeof M}:Fe=function(M){return M&&typeof Symbol=="function"&&M.constructor===Symbol&&M!==Symbol.prototype?"symbol":typeof M},Fe(V)}function ki(V,A){if(!(V instanceof A))throw new TypeError("Cannot call a class as a function")}function no(V,A){for(var M=0;M0&&arguments[0]!==void 0?arguments[0]:{};this.action=typeof D.action=="function"?D.action:this.defaultAction,this.target=typeof D.target=="function"?D.target:this.defaultTarget,this.text=typeof D.text=="function"?D.text:this.defaultText,this.container=Fe(D.container)==="object"?D.container:document.body}},{key:"listenClick",value:function(D){var Y=this;this.listener=c()(D,"click",function($e){return Y.onClick($e)})}},{key:"onClick",value:function(D){var Y=D.delegateTarget||D.currentTarget,$e=this.action(Y)||"copy",Dt=qe({action:$e,container:this.container,target:this.target(Y),text:this.text(Y)});this.emit(Dt?"success":"error",{action:$e,text:Dt,trigger:Y,clearSelection:function(){Y&&Y.focus(),window.getSelection().removeAllRanges()}})}},{key:"defaultAction",value:function(D){return vr("action",D)}},{key:"defaultTarget",value:function(D){var Y=vr("target",D);if(Y)return document.querySelector(Y)}},{key:"defaultText",value:function(D){return vr("text",D)}},{key:"destroy",value:function(){this.listener.destroy()}}],[{key:"copy",value:function(D){var Y=arguments.length>1&&arguments[1]!==void 0?arguments[1]:{container:document.body};return J(D,Y)}},{key:"cut",value:function(D){return y(D)}},{key:"isSupported",value:function(){var D=arguments.length>0&&arguments[0]!==void 0?arguments[0]:["copy","cut"],Y=typeof D=="string"?[D]:D,$e=!!document.queryCommandSupported;return Y.forEach(function(Dt){$e=$e&&!!document.queryCommandSupported(Dt)}),$e}}]),M}(s()),Ui=Fi},828:function(o){var n=9;if(typeof Element!="undefined"&&!Element.prototype.matches){var i=Element.prototype;i.matches=i.matchesSelector||i.mozMatchesSelector||i.msMatchesSelector||i.oMatchesSelector||i.webkitMatchesSelector}function a(s,p){for(;s&&s.nodeType!==n;){if(typeof s.matches=="function"&&s.matches(p))return s;s=s.parentNode}}o.exports=a},438:function(o,n,i){var a=i(828);function s(l,f,u,d,y){var L=c.apply(this,arguments);return l.addEventListener(u,L,y),{destroy:function(){l.removeEventListener(u,L,y)}}}function p(l,f,u,d,y){return typeof l.addEventListener=="function"?s.apply(null,arguments):typeof u=="function"?s.bind(null,document).apply(null,arguments):(typeof l=="string"&&(l=document.querySelectorAll(l)),Array.prototype.map.call(l,function(L){return s(L,f,u,d,y)}))}function c(l,f,u,d){return function(y){y.delegateTarget=a(y.target,f),y.delegateTarget&&d.call(l,y)}}o.exports=p},879:function(o,n){n.node=function(i){return i!==void 0&&i instanceof HTMLElement&&i.nodeType===1},n.nodeList=function(i){var a=Object.prototype.toString.call(i);return i!==void 0&&(a==="[object NodeList]"||a==="[object HTMLCollection]")&&"length"in i&&(i.length===0||n.node(i[0]))},n.string=function(i){return typeof i=="string"||i instanceof String},n.fn=function(i){var a=Object.prototype.toString.call(i);return a==="[object Function]"}},370:function(o,n,i){var a=i(879),s=i(438);function p(u,d,y){if(!u&&!d&&!y)throw new Error("Missing required arguments");if(!a.string(d))throw new TypeError("Second argument must be a String");if(!a.fn(y))throw new TypeError("Third argument must be a Function");if(a.node(u))return c(u,d,y);if(a.nodeList(u))return l(u,d,y);if(a.string(u))return f(u,d,y);throw new TypeError("First argument must be a String, HTMLElement, HTMLCollection, or NodeList")}function c(u,d,y){return u.addEventListener(d,y),{destroy:function(){u.removeEventListener(d,y)}}}function l(u,d,y){return Array.prototype.forEach.call(u,function(L){L.addEventListener(d,y)}),{destroy:function(){Array.prototype.forEach.call(u,function(L){L.removeEventListener(d,y)})}}}function f(u,d,y){return s(document.body,u,d,y)}o.exports=p},817:function(o){function n(i){var a;if(i.nodeName==="SELECT")i.focus(),a=i.value;else if(i.nodeName==="INPUT"||i.nodeName==="TEXTAREA"){var s=i.hasAttribute("readonly");s||i.setAttribute("readonly",""),i.select(),i.setSelectionRange(0,i.value.length),s||i.removeAttribute("readonly"),a=i.value}else{i.hasAttribute("contenteditable")&&i.focus();var p=window.getSelection(),c=document.createRange();c.selectNodeContents(i),p.removeAllRanges(),p.addRange(c),a=p.toString()}return a}o.exports=n},279:function(o){function n(){}n.prototype={on:function(i,a,s){var p=this.e||(this.e={});return(p[i]||(p[i]=[])).push({fn:a,ctx:s}),this},once:function(i,a,s){var p=this;function c(){p.off(i,c),a.apply(s,arguments)}return c._=a,this.on(i,c,s)},emit:function(i){var a=[].slice.call(arguments,1),s=((this.e||(this.e={}))[i]||[]).slice(),p=0,c=s.length;for(p;p0&&i[i.length-1])&&(c[0]===6||c[0]===2)){r=0;continue}if(c[0]===3&&(!i||c[1]>i[0]&&c[1]=e.length&&(e=void 0),{value:e&&e[o++],done:!e}}};throw new TypeError(t?"Object is not iterable.":"Symbol.iterator is not defined.")}function N(e,t){var r=typeof Symbol=="function"&&e[Symbol.iterator];if(!r)return e;var o=r.call(e),n,i=[],a;try{for(;(t===void 0||t-- >0)&&!(n=o.next()).done;)i.push(n.value)}catch(s){a={error:s}}finally{try{n&&!n.done&&(r=o.return)&&r.call(o)}finally{if(a)throw a.error}}return i}function q(e,t,r){if(r||arguments.length===2)for(var o=0,n=t.length,i;o1||p(d,L)})},y&&(n[d]=y(n[d])))}function p(d,y){try{c(o[d](y))}catch(L){u(i[0][3],L)}}function c(d){d.value instanceof nt?Promise.resolve(d.value.v).then(l,f):u(i[0][2],d)}function l(d){p("next",d)}function f(d){p("throw",d)}function u(d,y){d(y),i.shift(),i.length&&p(i[0][0],i[0][1])}}function uo(e){if(!Symbol.asyncIterator)throw new TypeError("Symbol.asyncIterator is not defined.");var t=e[Symbol.asyncIterator],r;return t?t.call(e):(e=typeof he=="function"?he(e):e[Symbol.iterator](),r={},o("next"),o("throw"),o("return"),r[Symbol.asyncIterator]=function(){return this},r);function o(i){r[i]=e[i]&&function(a){return new Promise(function(s,p){a=e[i](a),n(s,p,a.done,a.value)})}}function n(i,a,s,p){Promise.resolve(p).then(function(c){i({value:c,done:s})},a)}}function H(e){return typeof e=="function"}function ut(e){var t=function(o){Error.call(o),o.stack=new Error().stack},r=e(t);return r.prototype=Object.create(Error.prototype),r.prototype.constructor=r,r}var zt=ut(function(e){return function(r){e(this),this.message=r?r.length+` errors occurred during unsubscription: +`+r.map(function(o,n){return n+1+") "+o.toString()}).join(` + `):"",this.name="UnsubscriptionError",this.errors=r}});function Qe(e,t){if(e){var r=e.indexOf(t);0<=r&&e.splice(r,1)}}var Ue=function(){function e(t){this.initialTeardown=t,this.closed=!1,this._parentage=null,this._finalizers=null}return e.prototype.unsubscribe=function(){var t,r,o,n,i;if(!this.closed){this.closed=!0;var a=this._parentage;if(a)if(this._parentage=null,Array.isArray(a))try{for(var s=he(a),p=s.next();!p.done;p=s.next()){var c=p.value;c.remove(this)}}catch(L){t={error:L}}finally{try{p&&!p.done&&(r=s.return)&&r.call(s)}finally{if(t)throw t.error}}else a.remove(this);var l=this.initialTeardown;if(H(l))try{l()}catch(L){i=L instanceof zt?L.errors:[L]}var f=this._finalizers;if(f){this._finalizers=null;try{for(var u=he(f),d=u.next();!d.done;d=u.next()){var y=d.value;try{ho(y)}catch(L){i=i!=null?i:[],L instanceof zt?i=q(q([],N(i)),N(L.errors)):i.push(L)}}}catch(L){o={error:L}}finally{try{d&&!d.done&&(n=u.return)&&n.call(u)}finally{if(o)throw o.error}}}if(i)throw new zt(i)}},e.prototype.add=function(t){var r;if(t&&t!==this)if(this.closed)ho(t);else{if(t instanceof e){if(t.closed||t._hasParent(this))return;t._addParent(this)}(this._finalizers=(r=this._finalizers)!==null&&r!==void 0?r:[]).push(t)}},e.prototype._hasParent=function(t){var r=this._parentage;return r===t||Array.isArray(r)&&r.includes(t)},e.prototype._addParent=function(t){var r=this._parentage;this._parentage=Array.isArray(r)?(r.push(t),r):r?[r,t]:t},e.prototype._removeParent=function(t){var r=this._parentage;r===t?this._parentage=null:Array.isArray(r)&&Qe(r,t)},e.prototype.remove=function(t){var r=this._finalizers;r&&Qe(r,t),t instanceof e&&t._removeParent(this)},e.EMPTY=function(){var t=new e;return t.closed=!0,t}(),e}();var Tr=Ue.EMPTY;function qt(e){return e instanceof Ue||e&&"closed"in e&&H(e.remove)&&H(e.add)&&H(e.unsubscribe)}function ho(e){H(e)?e():e.unsubscribe()}var Pe={onUnhandledError:null,onStoppedNotification:null,Promise:void 0,useDeprecatedSynchronousErrorHandling:!1,useDeprecatedNextContext:!1};var dt={setTimeout:function(e,t){for(var r=[],o=2;o0},enumerable:!1,configurable:!0}),t.prototype._trySubscribe=function(r){return this._throwIfClosed(),e.prototype._trySubscribe.call(this,r)},t.prototype._subscribe=function(r){return this._throwIfClosed(),this._checkFinalizedStatuses(r),this._innerSubscribe(r)},t.prototype._innerSubscribe=function(r){var o=this,n=this,i=n.hasError,a=n.isStopped,s=n.observers;return i||a?Tr:(this.currentObservers=null,s.push(r),new Ue(function(){o.currentObservers=null,Qe(s,r)}))},t.prototype._checkFinalizedStatuses=function(r){var o=this,n=o.hasError,i=o.thrownError,a=o.isStopped;n?r.error(i):a&&r.complete()},t.prototype.asObservable=function(){var r=new j;return r.source=this,r},t.create=function(r,o){return new To(r,o)},t}(j);var To=function(e){oe(t,e);function t(r,o){var n=e.call(this)||this;return n.destination=r,n.source=o,n}return t.prototype.next=function(r){var o,n;(n=(o=this.destination)===null||o===void 0?void 0:o.next)===null||n===void 0||n.call(o,r)},t.prototype.error=function(r){var o,n;(n=(o=this.destination)===null||o===void 0?void 0:o.error)===null||n===void 0||n.call(o,r)},t.prototype.complete=function(){var r,o;(o=(r=this.destination)===null||r===void 0?void 0:r.complete)===null||o===void 0||o.call(r)},t.prototype._subscribe=function(r){var o,n;return(n=(o=this.source)===null||o===void 0?void 0:o.subscribe(r))!==null&&n!==void 0?n:Tr},t}(g);var _r=function(e){oe(t,e);function t(r){var o=e.call(this)||this;return o._value=r,o}return Object.defineProperty(t.prototype,"value",{get:function(){return this.getValue()},enumerable:!1,configurable:!0}),t.prototype._subscribe=function(r){var o=e.prototype._subscribe.call(this,r);return!o.closed&&r.next(this._value),o},t.prototype.getValue=function(){var r=this,o=r.hasError,n=r.thrownError,i=r._value;if(o)throw n;return this._throwIfClosed(),i},t.prototype.next=function(r){e.prototype.next.call(this,this._value=r)},t}(g);var At={now:function(){return(At.delegate||Date).now()},delegate:void 0};var Ct=function(e){oe(t,e);function t(r,o,n){r===void 0&&(r=1/0),o===void 0&&(o=1/0),n===void 0&&(n=At);var i=e.call(this)||this;return i._bufferSize=r,i._windowTime=o,i._timestampProvider=n,i._buffer=[],i._infiniteTimeWindow=!0,i._infiniteTimeWindow=o===1/0,i._bufferSize=Math.max(1,r),i._windowTime=Math.max(1,o),i}return t.prototype.next=function(r){var o=this,n=o.isStopped,i=o._buffer,a=o._infiniteTimeWindow,s=o._timestampProvider,p=o._windowTime;n||(i.push(r),!a&&i.push(s.now()+p)),this._trimBuffer(),e.prototype.next.call(this,r)},t.prototype._subscribe=function(r){this._throwIfClosed(),this._trimBuffer();for(var o=this._innerSubscribe(r),n=this,i=n._infiniteTimeWindow,a=n._buffer,s=a.slice(),p=0;p0?e.prototype.schedule.call(this,r,o):(this.delay=o,this.state=r,this.scheduler.flush(this),this)},t.prototype.execute=function(r,o){return o>0||this.closed?e.prototype.execute.call(this,r,o):this._execute(r,o)},t.prototype.requestAsyncId=function(r,o,n){return n===void 0&&(n=0),n!=null&&n>0||n==null&&this.delay>0?e.prototype.requestAsyncId.call(this,r,o,n):(r.flush(this),0)},t}(gt);var Lo=function(e){oe(t,e);function t(){return e!==null&&e.apply(this,arguments)||this}return t}(yt);var kr=new Lo(Oo);var Mo=function(e){oe(t,e);function t(r,o){var n=e.call(this,r,o)||this;return n.scheduler=r,n.work=o,n}return t.prototype.requestAsyncId=function(r,o,n){return n===void 0&&(n=0),n!==null&&n>0?e.prototype.requestAsyncId.call(this,r,o,n):(r.actions.push(this),r._scheduled||(r._scheduled=vt.requestAnimationFrame(function(){return r.flush(void 0)})))},t.prototype.recycleAsyncId=function(r,o,n){var i;if(n===void 0&&(n=0),n!=null?n>0:this.delay>0)return e.prototype.recycleAsyncId.call(this,r,o,n);var a=r.actions;o!=null&&((i=a[a.length-1])===null||i===void 0?void 0:i.id)!==o&&(vt.cancelAnimationFrame(o),r._scheduled=void 0)},t}(gt);var _o=function(e){oe(t,e);function t(){return e!==null&&e.apply(this,arguments)||this}return t.prototype.flush=function(r){this._active=!0;var o=this._scheduled;this._scheduled=void 0;var n=this.actions,i;r=r||n.shift();do if(i=r.execute(r.state,r.delay))break;while((r=n[0])&&r.id===o&&n.shift());if(this._active=!1,i){for(;(r=n[0])&&r.id===o&&n.shift();)r.unsubscribe();throw i}},t}(yt);var me=new _o(Mo);var S=new j(function(e){return e.complete()});function Yt(e){return e&&H(e.schedule)}function Hr(e){return e[e.length-1]}function Xe(e){return H(Hr(e))?e.pop():void 0}function ke(e){return Yt(Hr(e))?e.pop():void 0}function Bt(e,t){return typeof Hr(e)=="number"?e.pop():t}var xt=function(e){return e&&typeof e.length=="number"&&typeof e!="function"};function Gt(e){return H(e==null?void 0:e.then)}function Jt(e){return H(e[bt])}function Xt(e){return Symbol.asyncIterator&&H(e==null?void 0:e[Symbol.asyncIterator])}function Zt(e){return new TypeError("You provided "+(e!==null&&typeof e=="object"?"an invalid object":"'"+e+"'")+" where a stream was expected. You can provide an Observable, Promise, ReadableStream, Array, AsyncIterable, or Iterable.")}function Zi(){return typeof Symbol!="function"||!Symbol.iterator?"@@iterator":Symbol.iterator}var er=Zi();function tr(e){return H(e==null?void 0:e[er])}function rr(e){return fo(this,arguments,function(){var r,o,n,i;return Nt(this,function(a){switch(a.label){case 0:r=e.getReader(),a.label=1;case 1:a.trys.push([1,,9,10]),a.label=2;case 2:return[4,nt(r.read())];case 3:return o=a.sent(),n=o.value,i=o.done,i?[4,nt(void 0)]:[3,5];case 4:return[2,a.sent()];case 5:return[4,nt(n)];case 6:return[4,a.sent()];case 7:return a.sent(),[3,2];case 8:return[3,10];case 9:return r.releaseLock(),[7];case 10:return[2]}})})}function or(e){return H(e==null?void 0:e.getReader)}function U(e){if(e instanceof j)return e;if(e!=null){if(Jt(e))return ea(e);if(xt(e))return ta(e);if(Gt(e))return ra(e);if(Xt(e))return Ao(e);if(tr(e))return oa(e);if(or(e))return na(e)}throw Zt(e)}function ea(e){return new j(function(t){var r=e[bt]();if(H(r.subscribe))return r.subscribe(t);throw new TypeError("Provided object does not correctly implement Symbol.observable")})}function ta(e){return new j(function(t){for(var r=0;r=2;return function(o){return o.pipe(e?b(function(n,i){return e(n,i,o)}):le,Te(1),r?De(t):Qo(function(){return new ir}))}}function jr(e){return e<=0?function(){return S}:E(function(t,r){var o=[];t.subscribe(T(r,function(n){o.push(n),e=2,!0))}function pe(e){e===void 0&&(e={});var t=e.connector,r=t===void 0?function(){return new g}:t,o=e.resetOnError,n=o===void 0?!0:o,i=e.resetOnComplete,a=i===void 0?!0:i,s=e.resetOnRefCountZero,p=s===void 0?!0:s;return function(c){var l,f,u,d=0,y=!1,L=!1,X=function(){f==null||f.unsubscribe(),f=void 0},te=function(){X(),l=u=void 0,y=L=!1},J=function(){var k=l;te(),k==null||k.unsubscribe()};return E(function(k,ft){d++,!L&&!y&&X();var qe=u=u!=null?u:r();ft.add(function(){d--,d===0&&!L&&!y&&(f=Ur(J,p))}),qe.subscribe(ft),!l&&d>0&&(l=new at({next:function(Fe){return qe.next(Fe)},error:function(Fe){L=!0,X(),f=Ur(te,n,Fe),qe.error(Fe)},complete:function(){y=!0,X(),f=Ur(te,a),qe.complete()}}),U(k).subscribe(l))})(c)}}function Ur(e,t){for(var r=[],o=2;oe.next(document)),e}function P(e,t=document){return Array.from(t.querySelectorAll(e))}function R(e,t=document){let r=fe(e,t);if(typeof r=="undefined")throw new ReferenceError(`Missing element: expected "${e}" to be present`);return r}function fe(e,t=document){return t.querySelector(e)||void 0}function Ie(){var e,t,r,o;return(o=(r=(t=(e=document.activeElement)==null?void 0:e.shadowRoot)==null?void 0:t.activeElement)!=null?r:document.activeElement)!=null?o:void 0}var wa=O(h(document.body,"focusin"),h(document.body,"focusout")).pipe(_e(1),Q(void 0),m(()=>Ie()||document.body),G(1));function et(e){return wa.pipe(m(t=>e.contains(t)),K())}function $t(e,t){return C(()=>O(h(e,"mouseenter").pipe(m(()=>!0)),h(e,"mouseleave").pipe(m(()=>!1))).pipe(t?Ht(r=>Le(+!r*t)):le,Q(e.matches(":hover"))))}function Jo(e,t){if(typeof t=="string"||typeof t=="number")e.innerHTML+=t.toString();else if(t instanceof Node)e.appendChild(t);else if(Array.isArray(t))for(let r of t)Jo(e,r)}function x(e,t,...r){let o=document.createElement(e);if(t)for(let n of Object.keys(t))typeof t[n]!="undefined"&&(typeof t[n]!="boolean"?o.setAttribute(n,t[n]):o.setAttribute(n,""));for(let n of r)Jo(o,n);return o}function sr(e){if(e>999){let t=+((e-950)%1e3>99);return`${((e+1e-6)/1e3).toFixed(t)}k`}else return e.toString()}function Tt(e){let t=x("script",{src:e});return C(()=>(document.head.appendChild(t),O(h(t,"load"),h(t,"error").pipe(v(()=>$r(()=>new ReferenceError(`Invalid script: ${e}`))))).pipe(m(()=>{}),_(()=>document.head.removeChild(t)),Te(1))))}var Xo=new g,Ta=C(()=>typeof ResizeObserver=="undefined"?Tt("https://unpkg.com/resize-observer-polyfill"):I(void 0)).pipe(m(()=>new ResizeObserver(e=>e.forEach(t=>Xo.next(t)))),v(e=>O(Ye,I(e)).pipe(_(()=>e.disconnect()))),G(1));function ce(e){return{width:e.offsetWidth,height:e.offsetHeight}}function ge(e){let t=e;for(;t.clientWidth===0&&t.parentElement;)t=t.parentElement;return Ta.pipe(w(r=>r.observe(t)),v(r=>Xo.pipe(b(o=>o.target===t),_(()=>r.unobserve(t)))),m(()=>ce(e)),Q(ce(e)))}function St(e){return{width:e.scrollWidth,height:e.scrollHeight}}function cr(e){let t=e.parentElement;for(;t&&(e.scrollWidth<=t.scrollWidth&&e.scrollHeight<=t.scrollHeight);)t=(e=t).parentElement;return t?e:void 0}function Zo(e){let t=[],r=e.parentElement;for(;r;)(e.clientWidth>r.clientWidth||e.clientHeight>r.clientHeight)&&t.push(r),r=(e=r).parentElement;return t.length===0&&t.push(document.documentElement),t}function Ve(e){return{x:e.offsetLeft,y:e.offsetTop}}function en(e){let t=e.getBoundingClientRect();return{x:t.x+window.scrollX,y:t.y+window.scrollY}}function tn(e){return O(h(window,"load"),h(window,"resize")).pipe(Me(0,me),m(()=>Ve(e)),Q(Ve(e)))}function pr(e){return{x:e.scrollLeft,y:e.scrollTop}}function Ne(e){return O(h(e,"scroll"),h(window,"scroll"),h(window,"resize")).pipe(Me(0,me),m(()=>pr(e)),Q(pr(e)))}var rn=new g,Sa=C(()=>I(new IntersectionObserver(e=>{for(let t of e)rn.next(t)},{threshold:0}))).pipe(v(e=>O(Ye,I(e)).pipe(_(()=>e.disconnect()))),G(1));function tt(e){return Sa.pipe(w(t=>t.observe(e)),v(t=>rn.pipe(b(({target:r})=>r===e),_(()=>t.unobserve(e)),m(({isIntersecting:r})=>r))))}function on(e,t=16){return Ne(e).pipe(m(({y:r})=>{let o=ce(e),n=St(e);return r>=n.height-o.height-t}),K())}var lr={drawer:R("[data-md-toggle=drawer]"),search:R("[data-md-toggle=search]")};function nn(e){return lr[e].checked}function Je(e,t){lr[e].checked!==t&&lr[e].click()}function ze(e){let t=lr[e];return h(t,"change").pipe(m(()=>t.checked),Q(t.checked))}function Oa(e,t){switch(e.constructor){case HTMLInputElement:return e.type==="radio"?/^Arrow/.test(t):!0;case HTMLSelectElement:case HTMLTextAreaElement:return!0;default:return e.isContentEditable}}function La(){return O(h(window,"compositionstart").pipe(m(()=>!0)),h(window,"compositionend").pipe(m(()=>!1))).pipe(Q(!1))}function an(){let e=h(window,"keydown").pipe(b(t=>!(t.metaKey||t.ctrlKey)),m(t=>({mode:nn("search")?"search":"global",type:t.key,claim(){t.preventDefault(),t.stopPropagation()}})),b(({mode:t,type:r})=>{if(t==="global"){let o=Ie();if(typeof o!="undefined")return!Oa(o,r)}return!0}),pe());return La().pipe(v(t=>t?S:e))}function ye(){return new URL(location.href)}function lt(e,t=!1){if(B("navigation.instant")&&!t){let r=x("a",{href:e.href});document.body.appendChild(r),r.click(),r.remove()}else location.href=e.href}function sn(){return new g}function cn(){return location.hash.slice(1)}function pn(e){let t=x("a",{href:e});t.addEventListener("click",r=>r.stopPropagation()),t.click()}function Ma(e){return O(h(window,"hashchange"),e).pipe(m(cn),Q(cn()),b(t=>t.length>0),G(1))}function ln(e){return Ma(e).pipe(m(t=>fe(`[id="${t}"]`)),b(t=>typeof t!="undefined"))}function Pt(e){let t=matchMedia(e);return ar(r=>t.addListener(()=>r(t.matches))).pipe(Q(t.matches))}function mn(){let e=matchMedia("print");return O(h(window,"beforeprint").pipe(m(()=>!0)),h(window,"afterprint").pipe(m(()=>!1))).pipe(Q(e.matches))}function Nr(e,t){return e.pipe(v(r=>r?t():S))}function zr(e,t){return new j(r=>{let o=new XMLHttpRequest;return o.open("GET",`${e}`),o.responseType="blob",o.addEventListener("load",()=>{o.status>=200&&o.status<300?(r.next(o.response),r.complete()):r.error(new Error(o.statusText))}),o.addEventListener("error",()=>{r.error(new Error("Network error"))}),o.addEventListener("abort",()=>{r.complete()}),typeof(t==null?void 0:t.progress$)!="undefined"&&(o.addEventListener("progress",n=>{var i;if(n.lengthComputable)t.progress$.next(n.loaded/n.total*100);else{let a=(i=o.getResponseHeader("Content-Length"))!=null?i:0;t.progress$.next(n.loaded/+a*100)}}),t.progress$.next(5)),o.send(),()=>o.abort()})}function je(e,t){return zr(e,t).pipe(v(r=>r.text()),m(r=>JSON.parse(r)),G(1))}function fn(e,t){let r=new DOMParser;return zr(e,t).pipe(v(o=>o.text()),m(o=>r.parseFromString(o,"text/html")),G(1))}function un(e,t){let r=new DOMParser;return zr(e,t).pipe(v(o=>o.text()),m(o=>r.parseFromString(o,"text/xml")),G(1))}function dn(){return{x:Math.max(0,scrollX),y:Math.max(0,scrollY)}}function hn(){return O(h(window,"scroll",{passive:!0}),h(window,"resize",{passive:!0})).pipe(m(dn),Q(dn()))}function bn(){return{width:innerWidth,height:innerHeight}}function vn(){return h(window,"resize",{passive:!0}).pipe(m(bn),Q(bn()))}function gn(){return z([hn(),vn()]).pipe(m(([e,t])=>({offset:e,size:t})),G(1))}function mr(e,{viewport$:t,header$:r}){let o=t.pipe(ee("size")),n=z([o,r]).pipe(m(()=>Ve(e)));return z([r,t,n]).pipe(m(([{height:i},{offset:a,size:s},{x:p,y:c}])=>({offset:{x:a.x-p,y:a.y-c+i},size:s})))}function _a(e){return h(e,"message",t=>t.data)}function Aa(e){let t=new g;return t.subscribe(r=>e.postMessage(r)),t}function yn(e,t=new Worker(e)){let r=_a(t),o=Aa(t),n=new g;n.subscribe(o);let i=o.pipe(Z(),ie(!0));return n.pipe(Z(),Re(r.pipe(W(i))),pe())}var Ca=R("#__config"),Ot=JSON.parse(Ca.textContent);Ot.base=`${new URL(Ot.base,ye())}`;function xe(){return Ot}function B(e){return Ot.features.includes(e)}function Ee(e,t){return typeof t!="undefined"?Ot.translations[e].replace("#",t.toString()):Ot.translations[e]}function Se(e,t=document){return R(`[data-md-component=${e}]`,t)}function ae(e,t=document){return P(`[data-md-component=${e}]`,t)}function ka(e){let t=R(".md-typeset > :first-child",e);return h(t,"click",{once:!0}).pipe(m(()=>R(".md-typeset",e)),m(r=>({hash:__md_hash(r.innerHTML)})))}function xn(e){if(!B("announce.dismiss")||!e.childElementCount)return S;if(!e.hidden){let t=R(".md-typeset",e);__md_hash(t.innerHTML)===__md_get("__announce")&&(e.hidden=!0)}return C(()=>{let t=new g;return t.subscribe(({hash:r})=>{e.hidden=!0,__md_set("__announce",r)}),ka(e).pipe(w(r=>t.next(r)),_(()=>t.complete()),m(r=>$({ref:e},r)))})}function Ha(e,{target$:t}){return t.pipe(m(r=>({hidden:r!==e})))}function En(e,t){let r=new g;return r.subscribe(({hidden:o})=>{e.hidden=o}),Ha(e,t).pipe(w(o=>r.next(o)),_(()=>r.complete()),m(o=>$({ref:e},o)))}function Rt(e,t){return t==="inline"?x("div",{class:"md-tooltip md-tooltip--inline",id:e,role:"tooltip"},x("div",{class:"md-tooltip__inner md-typeset"})):x("div",{class:"md-tooltip",id:e,role:"tooltip"},x("div",{class:"md-tooltip__inner md-typeset"}))}function wn(...e){return x("div",{class:"md-tooltip2",role:"tooltip"},x("div",{class:"md-tooltip2__inner md-typeset"},e))}function Tn(e,t){if(t=t?`${t}_annotation_${e}`:void 0,t){let r=t?`#${t}`:void 0;return x("aside",{class:"md-annotation",tabIndex:0},Rt(t),x("a",{href:r,class:"md-annotation__index",tabIndex:-1},x("span",{"data-md-annotation-id":e})))}else return x("aside",{class:"md-annotation",tabIndex:0},Rt(t),x("span",{class:"md-annotation__index",tabIndex:-1},x("span",{"data-md-annotation-id":e})))}function Sn(e){return x("button",{class:"md-clipboard md-icon",title:Ee("clipboard.copy"),"data-clipboard-target":`#${e} > code`})}var Ln=Mt(qr());function Qr(e,t){let r=t&2,o=t&1,n=Object.keys(e.terms).filter(p=>!e.terms[p]).reduce((p,c)=>[...p,x("del",null,(0,Ln.default)(c))," "],[]).slice(0,-1),i=xe(),a=new URL(e.location,i.base);B("search.highlight")&&a.searchParams.set("h",Object.entries(e.terms).filter(([,p])=>p).reduce((p,[c])=>`${p} ${c}`.trim(),""));let{tags:s}=xe();return x("a",{href:`${a}`,class:"md-search-result__link",tabIndex:-1},x("article",{class:"md-search-result__article md-typeset","data-md-score":e.score.toFixed(2)},r>0&&x("div",{class:"md-search-result__icon md-icon"}),r>0&&x("h1",null,e.title),r<=0&&x("h2",null,e.title),o>0&&e.text.length>0&&e.text,e.tags&&x("nav",{class:"md-tags"},e.tags.map(p=>{let c=s?p in s?`md-tag-icon md-tag--${s[p]}`:"md-tag-icon":"";return x("span",{class:`md-tag ${c}`},p)})),o>0&&n.length>0&&x("p",{class:"md-search-result__terms"},Ee("search.result.term.missing"),": ",...n)))}function Mn(e){let t=e[0].score,r=[...e],o=xe(),n=r.findIndex(l=>!`${new URL(l.location,o.base)}`.includes("#")),[i]=r.splice(n,1),a=r.findIndex(l=>l.scoreQr(l,1)),...p.length?[x("details",{class:"md-search-result__more"},x("summary",{tabIndex:-1},x("div",null,p.length>0&&p.length===1?Ee("search.result.more.one"):Ee("search.result.more.other",p.length))),...p.map(l=>Qr(l,1)))]:[]];return x("li",{class:"md-search-result__item"},c)}function _n(e){return x("ul",{class:"md-source__facts"},Object.entries(e).map(([t,r])=>x("li",{class:`md-source__fact md-source__fact--${t}`},typeof r=="number"?sr(r):r)))}function Kr(e){let t=`tabbed-control tabbed-control--${e}`;return x("div",{class:t,hidden:!0},x("button",{class:"tabbed-button",tabIndex:-1,"aria-hidden":"true"}))}function An(e){return x("div",{class:"md-typeset__scrollwrap"},x("div",{class:"md-typeset__table"},e))}function Ra(e){var o;let t=xe(),r=new URL(`../${e.version}/`,t.base);return x("li",{class:"md-version__item"},x("a",{href:`${r}`,class:"md-version__link"},e.title,((o=t.version)==null?void 0:o.alias)&&e.aliases.length>0&&x("span",{class:"md-version__alias"},e.aliases[0])))}function Cn(e,t){var o;let r=xe();return e=e.filter(n=>{var i;return!((i=n.properties)!=null&&i.hidden)}),x("div",{class:"md-version"},x("button",{class:"md-version__current","aria-label":Ee("select.version")},t.title,((o=r.version)==null?void 0:o.alias)&&t.aliases.length>0&&x("span",{class:"md-version__alias"},t.aliases[0])),x("ul",{class:"md-version__list"},e.map(Ra)))}var Ia=0;function ja(e){let t=z([et(e),$t(e)]).pipe(m(([o,n])=>o||n),K()),r=C(()=>Zo(e)).pipe(ne(Ne),pt(1),He(t),m(()=>en(e)));return t.pipe(Ae(o=>o),v(()=>z([t,r])),m(([o,n])=>({active:o,offset:n})),pe())}function Fa(e,t){let{content$:r,viewport$:o}=t,n=`__tooltip2_${Ia++}`;return C(()=>{let i=new g,a=new _r(!1);i.pipe(Z(),ie(!1)).subscribe(a);let s=a.pipe(Ht(c=>Le(+!c*250,kr)),K(),v(c=>c?r:S),w(c=>c.id=n),pe());z([i.pipe(m(({active:c})=>c)),s.pipe(v(c=>$t(c,250)),Q(!1))]).pipe(m(c=>c.some(l=>l))).subscribe(a);let p=a.pipe(b(c=>c),re(s,o),m(([c,l,{size:f}])=>{let u=e.getBoundingClientRect(),d=u.width/2;if(l.role==="tooltip")return{x:d,y:8+u.height};if(u.y>=f.height/2){let{height:y}=ce(l);return{x:d,y:-16-y}}else return{x:d,y:16+u.height}}));return z([s,i,p]).subscribe(([c,{offset:l},f])=>{c.style.setProperty("--md-tooltip-host-x",`${l.x}px`),c.style.setProperty("--md-tooltip-host-y",`${l.y}px`),c.style.setProperty("--md-tooltip-x",`${f.x}px`),c.style.setProperty("--md-tooltip-y",`${f.y}px`),c.classList.toggle("md-tooltip2--top",f.y<0),c.classList.toggle("md-tooltip2--bottom",f.y>=0)}),a.pipe(b(c=>c),re(s,(c,l)=>l),b(c=>c.role==="tooltip")).subscribe(c=>{let l=ce(R(":scope > *",c));c.style.setProperty("--md-tooltip-width",`${l.width}px`),c.style.setProperty("--md-tooltip-tail","0px")}),a.pipe(K(),ve(me),re(s)).subscribe(([c,l])=>{l.classList.toggle("md-tooltip2--active",c)}),z([a.pipe(b(c=>c)),s]).subscribe(([c,l])=>{l.role==="dialog"?(e.setAttribute("aria-controls",n),e.setAttribute("aria-haspopup","dialog")):e.setAttribute("aria-describedby",n)}),a.pipe(b(c=>!c)).subscribe(()=>{e.removeAttribute("aria-controls"),e.removeAttribute("aria-describedby"),e.removeAttribute("aria-haspopup")}),ja(e).pipe(w(c=>i.next(c)),_(()=>i.complete()),m(c=>$({ref:e},c)))})}function mt(e,{viewport$:t},r=document.body){return Fa(e,{content$:new j(o=>{let n=e.title,i=wn(n);return o.next(i),e.removeAttribute("title"),r.append(i),()=>{i.remove(),e.setAttribute("title",n)}}),viewport$:t})}function Ua(e,t){let r=C(()=>z([tn(e),Ne(t)])).pipe(m(([{x:o,y:n},i])=>{let{width:a,height:s}=ce(e);return{x:o-i.x+a/2,y:n-i.y+s/2}}));return et(e).pipe(v(o=>r.pipe(m(n=>({active:o,offset:n})),Te(+!o||1/0))))}function kn(e,t,{target$:r}){let[o,n]=Array.from(e.children);return C(()=>{let i=new g,a=i.pipe(Z(),ie(!0));return i.subscribe({next({offset:s}){e.style.setProperty("--md-tooltip-x",`${s.x}px`),e.style.setProperty("--md-tooltip-y",`${s.y}px`)},complete(){e.style.removeProperty("--md-tooltip-x"),e.style.removeProperty("--md-tooltip-y")}}),tt(e).pipe(W(a)).subscribe(s=>{e.toggleAttribute("data-md-visible",s)}),O(i.pipe(b(({active:s})=>s)),i.pipe(_e(250),b(({active:s})=>!s))).subscribe({next({active:s}){s?e.prepend(o):o.remove()},complete(){e.prepend(o)}}),i.pipe(Me(16,me)).subscribe(({active:s})=>{o.classList.toggle("md-tooltip--active",s)}),i.pipe(pt(125,me),b(()=>!!e.offsetParent),m(()=>e.offsetParent.getBoundingClientRect()),m(({x:s})=>s)).subscribe({next(s){s?e.style.setProperty("--md-tooltip-0",`${-s}px`):e.style.removeProperty("--md-tooltip-0")},complete(){e.style.removeProperty("--md-tooltip-0")}}),h(n,"click").pipe(W(a),b(s=>!(s.metaKey||s.ctrlKey))).subscribe(s=>{s.stopPropagation(),s.preventDefault()}),h(n,"mousedown").pipe(W(a),re(i)).subscribe(([s,{active:p}])=>{var c;if(s.button!==0||s.metaKey||s.ctrlKey)s.preventDefault();else if(p){s.preventDefault();let l=e.parentElement.closest(".md-annotation");l instanceof HTMLElement?l.focus():(c=Ie())==null||c.blur()}}),r.pipe(W(a),b(s=>s===o),Ge(125)).subscribe(()=>e.focus()),Ua(e,t).pipe(w(s=>i.next(s)),_(()=>i.complete()),m(s=>$({ref:e},s)))})}function Wa(e){return e.tagName==="CODE"?P(".c, .c1, .cm",e):[e]}function Da(e){let t=[];for(let r of Wa(e)){let o=[],n=document.createNodeIterator(r,NodeFilter.SHOW_TEXT);for(let i=n.nextNode();i;i=n.nextNode())o.push(i);for(let i of o){let a;for(;a=/(\(\d+\))(!)?/.exec(i.textContent);){let[,s,p]=a;if(typeof p=="undefined"){let c=i.splitText(a.index);i=c.splitText(s.length),t.push(c)}else{i.textContent=s,t.push(i);break}}}}return t}function Hn(e,t){t.append(...Array.from(e.childNodes))}function fr(e,t,{target$:r,print$:o}){let n=t.closest("[id]"),i=n==null?void 0:n.id,a=new Map;for(let s of Da(t)){let[,p]=s.textContent.match(/\((\d+)\)/);fe(`:scope > li:nth-child(${p})`,e)&&(a.set(p,Tn(p,i)),s.replaceWith(a.get(p)))}return a.size===0?S:C(()=>{let s=new g,p=s.pipe(Z(),ie(!0)),c=[];for(let[l,f]of a)c.push([R(".md-typeset",f),R(`:scope > li:nth-child(${l})`,e)]);return o.pipe(W(p)).subscribe(l=>{e.hidden=!l,e.classList.toggle("md-annotation-list",l);for(let[f,u]of c)l?Hn(f,u):Hn(u,f)}),O(...[...a].map(([,l])=>kn(l,t,{target$:r}))).pipe(_(()=>s.complete()),pe())})}function $n(e){if(e.nextElementSibling){let t=e.nextElementSibling;if(t.tagName==="OL")return t;if(t.tagName==="P"&&!t.children.length)return $n(t)}}function Pn(e,t){return C(()=>{let r=$n(e);return typeof r!="undefined"?fr(r,e,t):S})}var Rn=Mt(Br());var Va=0;function In(e){if(e.nextElementSibling){let t=e.nextElementSibling;if(t.tagName==="OL")return t;if(t.tagName==="P"&&!t.children.length)return In(t)}}function Na(e){return ge(e).pipe(m(({width:t})=>({scrollable:St(e).width>t})),ee("scrollable"))}function jn(e,t){let{matches:r}=matchMedia("(hover)"),o=C(()=>{let n=new g,i=n.pipe(jr(1));n.subscribe(({scrollable:c})=>{c&&r?e.setAttribute("tabindex","0"):e.removeAttribute("tabindex")});let a=[];if(Rn.default.isSupported()&&(e.closest(".copy")||B("content.code.copy")&&!e.closest(".no-copy"))){let c=e.closest("pre");c.id=`__code_${Va++}`;let l=Sn(c.id);c.insertBefore(l,e),B("content.tooltips")&&a.push(mt(l,{viewport$}))}let s=e.closest(".highlight");if(s instanceof HTMLElement){let c=In(s);if(typeof c!="undefined"&&(s.classList.contains("annotate")||B("content.code.annotate"))){let l=fr(c,e,t);a.push(ge(s).pipe(W(i),m(({width:f,height:u})=>f&&u),K(),v(f=>f?l:S)))}}return P(":scope > span[id]",e).length&&e.classList.add("md-code__content"),Na(e).pipe(w(c=>n.next(c)),_(()=>n.complete()),m(c=>$({ref:e},c)),Re(...a))});return B("content.lazy")?tt(e).pipe(b(n=>n),Te(1),v(()=>o)):o}function za(e,{target$:t,print$:r}){let o=!0;return O(t.pipe(m(n=>n.closest("details:not([open])")),b(n=>e===n),m(()=>({action:"open",reveal:!0}))),r.pipe(b(n=>n||!o),w(()=>o=e.open),m(n=>({action:n?"open":"close"}))))}function Fn(e,t){return C(()=>{let r=new g;return r.subscribe(({action:o,reveal:n})=>{e.toggleAttribute("open",o==="open"),n&&e.scrollIntoView()}),za(e,t).pipe(w(o=>r.next(o)),_(()=>r.complete()),m(o=>$({ref:e},o)))})}var Un=".node circle,.node ellipse,.node path,.node polygon,.node rect{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}marker{fill:var(--md-mermaid-edge-color)!important}.edgeLabel .label rect{fill:#0000}.label{color:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.label foreignObject{line-height:normal;overflow:visible}.label div .edgeLabel{color:var(--md-mermaid-label-fg-color)}.edgeLabel,.edgeLabel p,.label div .edgeLabel{background-color:var(--md-mermaid-label-bg-color)}.edgeLabel,.edgeLabel p{fill:var(--md-mermaid-label-bg-color);color:var(--md-mermaid-edge-color)}.edgePath .path,.flowchart-link{stroke:var(--md-mermaid-edge-color);stroke-width:.05rem}.edgePath .arrowheadPath{fill:var(--md-mermaid-edge-color);stroke:none}.cluster rect{fill:var(--md-default-fg-color--lightest);stroke:var(--md-default-fg-color--lighter)}.cluster span{color:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}g #flowchart-circleEnd,g #flowchart-circleStart,g #flowchart-crossEnd,g #flowchart-crossStart,g #flowchart-pointEnd,g #flowchart-pointStart{stroke:none}g.classGroup line,g.classGroup rect{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}g.classGroup text{fill:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.classLabel .box{fill:var(--md-mermaid-label-bg-color);background-color:var(--md-mermaid-label-bg-color);opacity:1}.classLabel .label{fill:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.node .divider{stroke:var(--md-mermaid-node-fg-color)}.relation{stroke:var(--md-mermaid-edge-color)}.cardinality{fill:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.cardinality text{fill:inherit!important}defs #classDiagram-compositionEnd,defs #classDiagram-compositionStart,defs #classDiagram-dependencyEnd,defs #classDiagram-dependencyStart,defs #classDiagram-extensionEnd,defs #classDiagram-extensionStart{fill:var(--md-mermaid-edge-color)!important;stroke:var(--md-mermaid-edge-color)!important}defs #classDiagram-aggregationEnd,defs #classDiagram-aggregationStart{fill:var(--md-mermaid-label-bg-color)!important;stroke:var(--md-mermaid-edge-color)!important}g.stateGroup rect{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}g.stateGroup .state-title{fill:var(--md-mermaid-label-fg-color)!important;font-family:var(--md-mermaid-font-family)}g.stateGroup .composit{fill:var(--md-mermaid-label-bg-color)}.nodeLabel,.nodeLabel p{color:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}a .nodeLabel{text-decoration:underline}.node circle.state-end,.node circle.state-start,.start-state{fill:var(--md-mermaid-edge-color);stroke:none}.end-state-inner,.end-state-outer{fill:var(--md-mermaid-edge-color)}.end-state-inner,.node circle.state-end{stroke:var(--md-mermaid-label-bg-color)}.transition{stroke:var(--md-mermaid-edge-color)}[id^=state-fork] rect,[id^=state-join] rect{fill:var(--md-mermaid-edge-color)!important;stroke:none!important}.statediagram-cluster.statediagram-cluster .inner{fill:var(--md-default-bg-color)}.statediagram-cluster rect{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}.statediagram-state rect.divider{fill:var(--md-default-fg-color--lightest);stroke:var(--md-default-fg-color--lighter)}defs #statediagram-barbEnd{stroke:var(--md-mermaid-edge-color)}.attributeBoxEven,.attributeBoxOdd{fill:var(--md-mermaid-node-bg-color);stroke:var(--md-mermaid-node-fg-color)}.entityBox{fill:var(--md-mermaid-label-bg-color);stroke:var(--md-mermaid-node-fg-color)}.entityLabel{fill:var(--md-mermaid-label-fg-color);font-family:var(--md-mermaid-font-family)}.relationshipLabelBox{fill:var(--md-mermaid-label-bg-color);fill-opacity:1;background-color:var(--md-mermaid-label-bg-color);opacity:1}.relationshipLabel{fill:var(--md-mermaid-label-fg-color)}.relationshipLine{stroke:var(--md-mermaid-edge-color)}defs #ONE_OR_MORE_END *,defs #ONE_OR_MORE_START *,defs #ONLY_ONE_END *,defs #ONLY_ONE_START *,defs #ZERO_OR_MORE_END *,defs #ZERO_OR_MORE_START *,defs #ZERO_OR_ONE_END *,defs #ZERO_OR_ONE_START *{stroke:var(--md-mermaid-edge-color)!important}defs #ZERO_OR_MORE_END circle,defs #ZERO_OR_MORE_START circle{fill:var(--md-mermaid-label-bg-color)}.actor{fill:var(--md-mermaid-sequence-actor-bg-color);stroke:var(--md-mermaid-sequence-actor-border-color)}text.actor>tspan{fill:var(--md-mermaid-sequence-actor-fg-color);font-family:var(--md-mermaid-font-family)}line{stroke:var(--md-mermaid-sequence-actor-line-color)}.actor-man circle,.actor-man line{fill:var(--md-mermaid-sequence-actorman-bg-color);stroke:var(--md-mermaid-sequence-actorman-line-color)}.messageLine0,.messageLine1{stroke:var(--md-mermaid-sequence-message-line-color)}.note{fill:var(--md-mermaid-sequence-note-bg-color);stroke:var(--md-mermaid-sequence-note-border-color)}.loopText,.loopText>tspan,.messageText,.noteText>tspan{stroke:none;font-family:var(--md-mermaid-font-family)!important}.messageText{fill:var(--md-mermaid-sequence-message-fg-color)}.loopText,.loopText>tspan{fill:var(--md-mermaid-sequence-loop-fg-color)}.noteText>tspan{fill:var(--md-mermaid-sequence-note-fg-color)}#arrowhead path{fill:var(--md-mermaid-sequence-message-line-color);stroke:none}.loopLine{fill:var(--md-mermaid-sequence-loop-bg-color);stroke:var(--md-mermaid-sequence-loop-border-color)}.labelBox{fill:var(--md-mermaid-sequence-label-bg-color);stroke:none}.labelText,.labelText>span{fill:var(--md-mermaid-sequence-label-fg-color);font-family:var(--md-mermaid-font-family)}.sequenceNumber{fill:var(--md-mermaid-sequence-number-fg-color)}rect.rect{fill:var(--md-mermaid-sequence-box-bg-color);stroke:none}rect.rect+text.text{fill:var(--md-mermaid-sequence-box-fg-color)}defs #sequencenumber{fill:var(--md-mermaid-sequence-number-bg-color)!important}";var Gr,Qa=0;function Ka(){return typeof mermaid=="undefined"||mermaid instanceof Element?Tt("https://unpkg.com/mermaid@11/dist/mermaid.min.js"):I(void 0)}function Wn(e){return e.classList.remove("mermaid"),Gr||(Gr=Ka().pipe(w(()=>mermaid.initialize({startOnLoad:!1,themeCSS:Un,sequence:{actorFontSize:"16px",messageFontSize:"16px",noteFontSize:"16px"}})),m(()=>{}),G(1))),Gr.subscribe(()=>co(this,null,function*(){e.classList.add("mermaid");let t=`__mermaid_${Qa++}`,r=x("div",{class:"mermaid"}),o=e.textContent,{svg:n,fn:i}=yield mermaid.render(t,o),a=r.attachShadow({mode:"closed"});a.innerHTML=n,e.replaceWith(r),i==null||i(a)})),Gr.pipe(m(()=>({ref:e})))}var Dn=x("table");function Vn(e){return e.replaceWith(Dn),Dn.replaceWith(An(e)),I({ref:e})}function Ya(e){let t=e.find(r=>r.checked)||e[0];return O(...e.map(r=>h(r,"change").pipe(m(()=>R(`label[for="${r.id}"]`))))).pipe(Q(R(`label[for="${t.id}"]`)),m(r=>({active:r})))}function Nn(e,{viewport$:t,target$:r}){let o=R(".tabbed-labels",e),n=P(":scope > input",e),i=Kr("prev");e.append(i);let a=Kr("next");return e.append(a),C(()=>{let s=new g,p=s.pipe(Z(),ie(!0));z([s,ge(e),tt(e)]).pipe(W(p),Me(1,me)).subscribe({next([{active:c},l]){let f=Ve(c),{width:u}=ce(c);e.style.setProperty("--md-indicator-x",`${f.x}px`),e.style.setProperty("--md-indicator-width",`${u}px`);let d=pr(o);(f.xd.x+l.width)&&o.scrollTo({left:Math.max(0,f.x-16),behavior:"smooth"})},complete(){e.style.removeProperty("--md-indicator-x"),e.style.removeProperty("--md-indicator-width")}}),z([Ne(o),ge(o)]).pipe(W(p)).subscribe(([c,l])=>{let f=St(o);i.hidden=c.x<16,a.hidden=c.x>f.width-l.width-16}),O(h(i,"click").pipe(m(()=>-1)),h(a,"click").pipe(m(()=>1))).pipe(W(p)).subscribe(c=>{let{width:l}=ce(o);o.scrollBy({left:l*c,behavior:"smooth"})}),r.pipe(W(p),b(c=>n.includes(c))).subscribe(c=>c.click()),o.classList.add("tabbed-labels--linked");for(let c of n){let l=R(`label[for="${c.id}"]`);l.replaceChildren(x("a",{href:`#${l.htmlFor}`,tabIndex:-1},...Array.from(l.childNodes))),h(l.firstElementChild,"click").pipe(W(p),b(f=>!(f.metaKey||f.ctrlKey)),w(f=>{f.preventDefault(),f.stopPropagation()})).subscribe(()=>{history.replaceState({},"",`#${l.htmlFor}`),l.click()})}return B("content.tabs.link")&&s.pipe(Ce(1),re(t)).subscribe(([{active:c},{offset:l}])=>{let f=c.innerText.trim();if(c.hasAttribute("data-md-switching"))c.removeAttribute("data-md-switching");else{let u=e.offsetTop-l.y;for(let y of P("[data-tabs]"))for(let L of P(":scope > input",y)){let X=R(`label[for="${L.id}"]`);if(X!==c&&X.innerText.trim()===f){X.setAttribute("data-md-switching",""),L.click();break}}window.scrollTo({top:e.offsetTop-u});let d=__md_get("__tabs")||[];__md_set("__tabs",[...new Set([f,...d])])}}),s.pipe(W(p)).subscribe(()=>{for(let c of P("audio, video",e))c.pause()}),Ya(n).pipe(w(c=>s.next(c)),_(()=>s.complete()),m(c=>$({ref:e},c)))}).pipe(Ke(se))}function zn(e,{viewport$:t,target$:r,print$:o}){return O(...P(".annotate:not(.highlight)",e).map(n=>Pn(n,{target$:r,print$:o})),...P("pre:not(.mermaid) > code",e).map(n=>jn(n,{target$:r,print$:o})),...P("pre.mermaid",e).map(n=>Wn(n)),...P("table:not([class])",e).map(n=>Vn(n)),...P("details",e).map(n=>Fn(n,{target$:r,print$:o})),...P("[data-tabs]",e).map(n=>Nn(n,{viewport$:t,target$:r})),...P("[title]",e).filter(()=>B("content.tooltips")).map(n=>mt(n,{viewport$:t})))}function Ba(e,{alert$:t}){return t.pipe(v(r=>O(I(!0),I(!1).pipe(Ge(2e3))).pipe(m(o=>({message:r,active:o})))))}function qn(e,t){let r=R(".md-typeset",e);return C(()=>{let o=new g;return o.subscribe(({message:n,active:i})=>{e.classList.toggle("md-dialog--active",i),r.textContent=n}),Ba(e,t).pipe(w(n=>o.next(n)),_(()=>o.complete()),m(n=>$({ref:e},n)))})}var Ga=0;function Ja(e,t){document.body.append(e);let{width:r}=ce(e);e.style.setProperty("--md-tooltip-width",`${r}px`),e.remove();let o=cr(t),n=typeof o!="undefined"?Ne(o):I({x:0,y:0}),i=O(et(t),$t(t)).pipe(K());return z([i,n]).pipe(m(([a,s])=>{let{x:p,y:c}=Ve(t),l=ce(t),f=t.closest("table");return f&&t.parentElement&&(p+=f.offsetLeft+t.parentElement.offsetLeft,c+=f.offsetTop+t.parentElement.offsetTop),{active:a,offset:{x:p-s.x+l.width/2-r/2,y:c-s.y+l.height+8}}}))}function Qn(e){let t=e.title;if(!t.length)return S;let r=`__tooltip_${Ga++}`,o=Rt(r,"inline"),n=R(".md-typeset",o);return n.innerHTML=t,C(()=>{let i=new g;return i.subscribe({next({offset:a}){o.style.setProperty("--md-tooltip-x",`${a.x}px`),o.style.setProperty("--md-tooltip-y",`${a.y}px`)},complete(){o.style.removeProperty("--md-tooltip-x"),o.style.removeProperty("--md-tooltip-y")}}),O(i.pipe(b(({active:a})=>a)),i.pipe(_e(250),b(({active:a})=>!a))).subscribe({next({active:a}){a?(e.insertAdjacentElement("afterend",o),e.setAttribute("aria-describedby",r),e.removeAttribute("title")):(o.remove(),e.removeAttribute("aria-describedby"),e.setAttribute("title",t))},complete(){o.remove(),e.removeAttribute("aria-describedby"),e.setAttribute("title",t)}}),i.pipe(Me(16,me)).subscribe(({active:a})=>{o.classList.toggle("md-tooltip--active",a)}),i.pipe(pt(125,me),b(()=>!!e.offsetParent),m(()=>e.offsetParent.getBoundingClientRect()),m(({x:a})=>a)).subscribe({next(a){a?o.style.setProperty("--md-tooltip-0",`${-a}px`):o.style.removeProperty("--md-tooltip-0")},complete(){o.style.removeProperty("--md-tooltip-0")}}),Ja(o,e).pipe(w(a=>i.next(a)),_(()=>i.complete()),m(a=>$({ref:e},a)))}).pipe(Ke(se))}function Xa({viewport$:e}){if(!B("header.autohide"))return I(!1);let t=e.pipe(m(({offset:{y:n}})=>n),Be(2,1),m(([n,i])=>[nMath.abs(i-n.y)>100),m(([,[n]])=>n),K()),o=ze("search");return z([e,o]).pipe(m(([{offset:n},i])=>n.y>400&&!i),K(),v(n=>n?r:I(!1)),Q(!1))}function Kn(e,t){return C(()=>z([ge(e),Xa(t)])).pipe(m(([{height:r},o])=>({height:r,hidden:o})),K((r,o)=>r.height===o.height&&r.hidden===o.hidden),G(1))}function Yn(e,{header$:t,main$:r}){return C(()=>{let o=new g,n=o.pipe(Z(),ie(!0));o.pipe(ee("active"),He(t)).subscribe(([{active:a},{hidden:s}])=>{e.classList.toggle("md-header--shadow",a&&!s),e.hidden=s});let i=ue(P("[title]",e)).pipe(b(()=>B("content.tooltips")),ne(a=>Qn(a)));return r.subscribe(o),t.pipe(W(n),m(a=>$({ref:e},a)),Re(i.pipe(W(n))))})}function Za(e,{viewport$:t,header$:r}){return mr(e,{viewport$:t,header$:r}).pipe(m(({offset:{y:o}})=>{let{height:n}=ce(e);return{active:o>=n}}),ee("active"))}function Bn(e,t){return C(()=>{let r=new g;r.subscribe({next({active:n}){e.classList.toggle("md-header__title--active",n)},complete(){e.classList.remove("md-header__title--active")}});let o=fe(".md-content h1");return typeof o=="undefined"?S:Za(o,t).pipe(w(n=>r.next(n)),_(()=>r.complete()),m(n=>$({ref:e},n)))})}function Gn(e,{viewport$:t,header$:r}){let o=r.pipe(m(({height:i})=>i),K()),n=o.pipe(v(()=>ge(e).pipe(m(({height:i})=>({top:e.offsetTop,bottom:e.offsetTop+i})),ee("bottom"))));return z([o,n,t]).pipe(m(([i,{top:a,bottom:s},{offset:{y:p},size:{height:c}}])=>(c=Math.max(0,c-Math.max(0,a-p,i)-Math.max(0,c+p-s)),{offset:a-i,height:c,active:a-i<=p})),K((i,a)=>i.offset===a.offset&&i.height===a.height&&i.active===a.active))}function es(e){let t=__md_get("__palette")||{index:e.findIndex(o=>matchMedia(o.getAttribute("data-md-color-media")).matches)},r=Math.max(0,Math.min(t.index,e.length-1));return I(...e).pipe(ne(o=>h(o,"change").pipe(m(()=>o))),Q(e[r]),m(o=>({index:e.indexOf(o),color:{media:o.getAttribute("data-md-color-media"),scheme:o.getAttribute("data-md-color-scheme"),primary:o.getAttribute("data-md-color-primary"),accent:o.getAttribute("data-md-color-accent")}})),G(1))}function Jn(e){let t=P("input",e),r=x("meta",{name:"theme-color"});document.head.appendChild(r);let o=x("meta",{name:"color-scheme"});document.head.appendChild(o);let n=Pt("(prefers-color-scheme: light)");return C(()=>{let i=new g;return i.subscribe(a=>{if(document.body.setAttribute("data-md-color-switching",""),a.color.media==="(prefers-color-scheme)"){let s=matchMedia("(prefers-color-scheme: light)"),p=document.querySelector(s.matches?"[data-md-color-media='(prefers-color-scheme: light)']":"[data-md-color-media='(prefers-color-scheme: dark)']");a.color.scheme=p.getAttribute("data-md-color-scheme"),a.color.primary=p.getAttribute("data-md-color-primary"),a.color.accent=p.getAttribute("data-md-color-accent")}for(let[s,p]of Object.entries(a.color))document.body.setAttribute(`data-md-color-${s}`,p);for(let s=0;sa.key==="Enter"),re(i,(a,s)=>s)).subscribe(({index:a})=>{a=(a+1)%t.length,t[a].click(),t[a].focus()}),i.pipe(m(()=>{let a=Se("header"),s=window.getComputedStyle(a);return o.content=s.colorScheme,s.backgroundColor.match(/\d+/g).map(p=>(+p).toString(16).padStart(2,"0")).join("")})).subscribe(a=>r.content=`#${a}`),i.pipe(ve(se)).subscribe(()=>{document.body.removeAttribute("data-md-color-switching")}),es(t).pipe(W(n.pipe(Ce(1))),ct(),w(a=>i.next(a)),_(()=>i.complete()),m(a=>$({ref:e},a)))})}function Xn(e,{progress$:t}){return C(()=>{let r=new g;return r.subscribe(({value:o})=>{e.style.setProperty("--md-progress-value",`${o}`)}),t.pipe(w(o=>r.next({value:o})),_(()=>r.complete()),m(o=>({ref:e,value:o})))})}var Jr=Mt(Br());function ts(e){e.setAttribute("data-md-copying","");let t=e.closest("[data-copy]"),r=t?t.getAttribute("data-copy"):e.innerText;return e.removeAttribute("data-md-copying"),r.trimEnd()}function Zn({alert$:e}){Jr.default.isSupported()&&new j(t=>{new Jr.default("[data-clipboard-target], [data-clipboard-text]",{text:r=>r.getAttribute("data-clipboard-text")||ts(R(r.getAttribute("data-clipboard-target")))}).on("success",r=>t.next(r))}).pipe(w(t=>{t.trigger.focus()}),m(()=>Ee("clipboard.copied"))).subscribe(e)}function ei(e,t){return e.protocol=t.protocol,e.hostname=t.hostname,e}function rs(e,t){let r=new Map;for(let o of P("url",e)){let n=R("loc",o),i=[ei(new URL(n.textContent),t)];r.set(`${i[0]}`,i);for(let a of P("[rel=alternate]",o)){let s=a.getAttribute("href");s!=null&&i.push(ei(new URL(s),t))}}return r}function ur(e){return un(new URL("sitemap.xml",e)).pipe(m(t=>rs(t,new URL(e))),de(()=>I(new Map)))}function os(e,t){if(!(e.target instanceof Element))return S;let r=e.target.closest("a");if(r===null)return S;if(r.target||e.metaKey||e.ctrlKey)return S;let o=new URL(r.href);return o.search=o.hash="",t.has(`${o}`)?(e.preventDefault(),I(new URL(r.href))):S}function ti(e){let t=new Map;for(let r of P(":scope > *",e.head))t.set(r.outerHTML,r);return t}function ri(e){for(let t of P("[href], [src]",e))for(let r of["href","src"]){let o=t.getAttribute(r);if(o&&!/^(?:[a-z]+:)?\/\//i.test(o)){t[r]=t[r];break}}return I(e)}function ns(e){for(let o of["[data-md-component=announce]","[data-md-component=container]","[data-md-component=header-topic]","[data-md-component=outdated]","[data-md-component=logo]","[data-md-component=skip]",...B("navigation.tabs.sticky")?["[data-md-component=tabs]"]:[]]){let n=fe(o),i=fe(o,e);typeof n!="undefined"&&typeof i!="undefined"&&n.replaceWith(i)}let t=ti(document);for(let[o,n]of ti(e))t.has(o)?t.delete(o):document.head.appendChild(n);for(let o of t.values()){let n=o.getAttribute("name");n!=="theme-color"&&n!=="color-scheme"&&o.remove()}let r=Se("container");return We(P("script",r)).pipe(v(o=>{let n=e.createElement("script");if(o.src){for(let i of o.getAttributeNames())n.setAttribute(i,o.getAttribute(i));return o.replaceWith(n),new j(i=>{n.onload=()=>i.complete()})}else return n.textContent=o.textContent,o.replaceWith(n),S}),Z(),ie(document))}function oi({location$:e,viewport$:t,progress$:r}){let o=xe();if(location.protocol==="file:")return S;let n=ur(o.base);I(document).subscribe(ri);let i=h(document.body,"click").pipe(He(n),v(([p,c])=>os(p,c)),pe()),a=h(window,"popstate").pipe(m(ye),pe());i.pipe(re(t)).subscribe(([p,{offset:c}])=>{history.replaceState(c,""),history.pushState(null,"",p)}),O(i,a).subscribe(e);let s=e.pipe(ee("pathname"),v(p=>fn(p,{progress$:r}).pipe(de(()=>(lt(p,!0),S)))),v(ri),v(ns),pe());return O(s.pipe(re(e,(p,c)=>c)),s.pipe(v(()=>e),ee("pathname"),v(()=>e),ee("hash")),e.pipe(K((p,c)=>p.pathname===c.pathname&&p.hash===c.hash),v(()=>i),w(()=>history.back()))).subscribe(p=>{var c,l;history.state!==null||!p.hash?window.scrollTo(0,(l=(c=history.state)==null?void 0:c.y)!=null?l:0):(history.scrollRestoration="auto",pn(p.hash),history.scrollRestoration="manual")}),e.subscribe(()=>{history.scrollRestoration="manual"}),h(window,"beforeunload").subscribe(()=>{history.scrollRestoration="auto"}),t.pipe(ee("offset"),_e(100)).subscribe(({offset:p})=>{history.replaceState(p,"")}),s}var ni=Mt(qr());function ii(e){let t=e.separator.split("|").map(n=>n.replace(/(\(\?[!=<][^)]+\))/g,"").length===0?"\uFFFD":n).join("|"),r=new RegExp(t,"img"),o=(n,i,a)=>`${i}${a}`;return n=>{n=n.replace(/[\s*+\-:~^]+/g," ").trim();let i=new RegExp(`(^|${e.separator}|)(${n.replace(/[|\\{}()[\]^$+*?.-]/g,"\\$&").replace(r,"|")})`,"img");return a=>(0,ni.default)(a).replace(i,o).replace(/<\/mark>(\s+)]*>/img,"$1")}}function jt(e){return e.type===1}function dr(e){return e.type===3}function ai(e,t){let r=yn(e);return O(I(location.protocol!=="file:"),ze("search")).pipe(Ae(o=>o),v(()=>t)).subscribe(({config:o,docs:n})=>r.next({type:0,data:{config:o,docs:n,options:{suggest:B("search.suggest")}}})),r}function si(e){var l;let{selectedVersionSitemap:t,selectedVersionBaseURL:r,currentLocation:o,currentBaseURL:n}=e,i=(l=Xr(n))==null?void 0:l.pathname;if(i===void 0)return;let a=ss(o.pathname,i);if(a===void 0)return;let s=ps(t.keys());if(!t.has(s))return;let p=Xr(a,s);if(!p||!t.has(p.href))return;let c=Xr(a,r);if(c)return c.hash=o.hash,c.search=o.search,c}function Xr(e,t){try{return new URL(e,t)}catch(r){return}}function ss(e,t){if(e.startsWith(t))return e.slice(t.length)}function cs(e,t){let r=Math.min(e.length,t.length),o;for(o=0;oS)),o=r.pipe(m(n=>{let[,i]=t.base.match(/([^/]+)\/?$/);return n.find(({version:a,aliases:s})=>a===i||s.includes(i))||n[0]}));r.pipe(m(n=>new Map(n.map(i=>[`${new URL(`../${i.version}/`,t.base)}`,i]))),v(n=>h(document.body,"click").pipe(b(i=>!i.metaKey&&!i.ctrlKey),re(o),v(([i,a])=>{if(i.target instanceof Element){let s=i.target.closest("a");if(s&&!s.target&&n.has(s.href)){let p=s.href;return!i.target.closest(".md-version")&&n.get(p)===a?S:(i.preventDefault(),I(new URL(p)))}}return S}),v(i=>ur(i).pipe(m(a=>{var s;return(s=si({selectedVersionSitemap:a,selectedVersionBaseURL:i,currentLocation:ye(),currentBaseURL:t.base}))!=null?s:i})))))).subscribe(n=>lt(n,!0)),z([r,o]).subscribe(([n,i])=>{R(".md-header__topic").appendChild(Cn(n,i))}),e.pipe(v(()=>o)).subscribe(n=>{var a;let i=__md_get("__outdated",sessionStorage);if(i===null){i=!0;let s=((a=t.version)==null?void 0:a.default)||"latest";Array.isArray(s)||(s=[s]);e:for(let p of s)for(let c of n.aliases.concat(n.version))if(new RegExp(p,"i").test(c)){i=!1;break e}__md_set("__outdated",i,sessionStorage)}if(i)for(let s of ae("outdated"))s.hidden=!1})}function ls(e,{worker$:t}){let{searchParams:r}=ye();r.has("q")&&(Je("search",!0),e.value=r.get("q"),e.focus(),ze("search").pipe(Ae(i=>!i)).subscribe(()=>{let i=ye();i.searchParams.delete("q"),history.replaceState({},"",`${i}`)}));let o=et(e),n=O(t.pipe(Ae(jt)),h(e,"keyup"),o).pipe(m(()=>e.value),K());return z([n,o]).pipe(m(([i,a])=>({value:i,focus:a})),G(1))}function pi(e,{worker$:t}){let r=new g,o=r.pipe(Z(),ie(!0));z([t.pipe(Ae(jt)),r],(i,a)=>a).pipe(ee("value")).subscribe(({value:i})=>t.next({type:2,data:i})),r.pipe(ee("focus")).subscribe(({focus:i})=>{i&&Je("search",i)}),h(e.form,"reset").pipe(W(o)).subscribe(()=>e.focus());let n=R("header [for=__search]");return h(n,"click").subscribe(()=>e.focus()),ls(e,{worker$:t}).pipe(w(i=>r.next(i)),_(()=>r.complete()),m(i=>$({ref:e},i)),G(1))}function li(e,{worker$:t,query$:r}){let o=new g,n=on(e.parentElement).pipe(b(Boolean)),i=e.parentElement,a=R(":scope > :first-child",e),s=R(":scope > :last-child",e);ze("search").subscribe(l=>s.setAttribute("role",l?"list":"presentation")),o.pipe(re(r),Wr(t.pipe(Ae(jt)))).subscribe(([{items:l},{value:f}])=>{switch(l.length){case 0:a.textContent=f.length?Ee("search.result.none"):Ee("search.result.placeholder");break;case 1:a.textContent=Ee("search.result.one");break;default:let u=sr(l.length);a.textContent=Ee("search.result.other",u)}});let p=o.pipe(w(()=>s.innerHTML=""),v(({items:l})=>O(I(...l.slice(0,10)),I(...l.slice(10)).pipe(Be(4),Vr(n),v(([f])=>f)))),m(Mn),pe());return p.subscribe(l=>s.appendChild(l)),p.pipe(ne(l=>{let f=fe("details",l);return typeof f=="undefined"?S:h(f,"toggle").pipe(W(o),m(()=>f))})).subscribe(l=>{l.open===!1&&l.offsetTop<=i.scrollTop&&i.scrollTo({top:l.offsetTop})}),t.pipe(b(dr),m(({data:l})=>l)).pipe(w(l=>o.next(l)),_(()=>o.complete()),m(l=>$({ref:e},l)))}function ms(e,{query$:t}){return t.pipe(m(({value:r})=>{let o=ye();return o.hash="",r=r.replace(/\s+/g,"+").replace(/&/g,"%26").replace(/=/g,"%3D"),o.search=`q=${r}`,{url:o}}))}function mi(e,t){let r=new g,o=r.pipe(Z(),ie(!0));return r.subscribe(({url:n})=>{e.setAttribute("data-clipboard-text",e.href),e.href=`${n}`}),h(e,"click").pipe(W(o)).subscribe(n=>n.preventDefault()),ms(e,t).pipe(w(n=>r.next(n)),_(()=>r.complete()),m(n=>$({ref:e},n)))}function fi(e,{worker$:t,keyboard$:r}){let o=new g,n=Se("search-query"),i=O(h(n,"keydown"),h(n,"focus")).pipe(ve(se),m(()=>n.value),K());return o.pipe(He(i),m(([{suggest:s},p])=>{let c=p.split(/([\s-]+)/);if(s!=null&&s.length&&c[c.length-1]){let l=s[s.length-1];l.startsWith(c[c.length-1])&&(c[c.length-1]=l)}else c.length=0;return c})).subscribe(s=>e.innerHTML=s.join("").replace(/\s/g," ")),r.pipe(b(({mode:s})=>s==="search")).subscribe(s=>{switch(s.type){case"ArrowRight":e.innerText.length&&n.selectionStart===n.value.length&&(n.value=e.innerText);break}}),t.pipe(b(dr),m(({data:s})=>s)).pipe(w(s=>o.next(s)),_(()=>o.complete()),m(()=>({ref:e})))}function ui(e,{index$:t,keyboard$:r}){let o=xe();try{let n=ai(o.search,t),i=Se("search-query",e),a=Se("search-result",e);h(e,"click").pipe(b(({target:p})=>p instanceof Element&&!!p.closest("a"))).subscribe(()=>Je("search",!1)),r.pipe(b(({mode:p})=>p==="search")).subscribe(p=>{let c=Ie();switch(p.type){case"Enter":if(c===i){let l=new Map;for(let f of P(":first-child [href]",a)){let u=f.firstElementChild;l.set(f,parseFloat(u.getAttribute("data-md-score")))}if(l.size){let[[f]]=[...l].sort(([,u],[,d])=>d-u);f.click()}p.claim()}break;case"Escape":case"Tab":Je("search",!1),i.blur();break;case"ArrowUp":case"ArrowDown":if(typeof c=="undefined")i.focus();else{let l=[i,...P(":not(details) > [href], summary, details[open] [href]",a)],f=Math.max(0,(Math.max(0,l.indexOf(c))+l.length+(p.type==="ArrowUp"?-1:1))%l.length);l[f].focus()}p.claim();break;default:i!==Ie()&&i.focus()}}),r.pipe(b(({mode:p})=>p==="global")).subscribe(p=>{switch(p.type){case"f":case"s":case"/":i.focus(),i.select(),p.claim();break}});let s=pi(i,{worker$:n});return O(s,li(a,{worker$:n,query$:s})).pipe(Re(...ae("search-share",e).map(p=>mi(p,{query$:s})),...ae("search-suggest",e).map(p=>fi(p,{worker$:n,keyboard$:r}))))}catch(n){return e.hidden=!0,Ye}}function di(e,{index$:t,location$:r}){return z([t,r.pipe(Q(ye()),b(o=>!!o.searchParams.get("h")))]).pipe(m(([o,n])=>ii(o.config)(n.searchParams.get("h"))),m(o=>{var a;let n=new Map,i=document.createNodeIterator(e,NodeFilter.SHOW_TEXT);for(let s=i.nextNode();s;s=i.nextNode())if((a=s.parentElement)!=null&&a.offsetHeight){let p=s.textContent,c=o(p);c.length>p.length&&n.set(s,c)}for(let[s,p]of n){let{childNodes:c}=x("span",null,p);s.replaceWith(...Array.from(c))}return{ref:e,nodes:n}}))}function fs(e,{viewport$:t,main$:r}){let o=e.closest(".md-grid"),n=o.offsetTop-o.parentElement.offsetTop;return z([r,t]).pipe(m(([{offset:i,height:a},{offset:{y:s}}])=>(a=a+Math.min(n,Math.max(0,s-i))-n,{height:a,locked:s>=i+n})),K((i,a)=>i.height===a.height&&i.locked===a.locked))}function Zr(e,o){var n=o,{header$:t}=n,r=so(n,["header$"]);let i=R(".md-sidebar__scrollwrap",e),{y:a}=Ve(i);return C(()=>{let s=new g,p=s.pipe(Z(),ie(!0)),c=s.pipe(Me(0,me));return c.pipe(re(t)).subscribe({next([{height:l},{height:f}]){i.style.height=`${l-2*a}px`,e.style.top=`${f}px`},complete(){i.style.height="",e.style.top=""}}),c.pipe(Ae()).subscribe(()=>{for(let l of P(".md-nav__link--active[href]",e)){if(!l.clientHeight)continue;let f=l.closest(".md-sidebar__scrollwrap");if(typeof f!="undefined"){let u=l.offsetTop-f.offsetTop,{height:d}=ce(f);f.scrollTo({top:u-d/2})}}}),ue(P("label[tabindex]",e)).pipe(ne(l=>h(l,"click").pipe(ve(se),m(()=>l),W(p)))).subscribe(l=>{let f=R(`[id="${l.htmlFor}"]`);R(`[aria-labelledby="${l.id}"]`).setAttribute("aria-expanded",`${f.checked}`)}),fs(e,r).pipe(w(l=>s.next(l)),_(()=>s.complete()),m(l=>$({ref:e},l)))})}function hi(e,t){if(typeof t!="undefined"){let r=`https://api.github.com/repos/${e}/${t}`;return st(je(`${r}/releases/latest`).pipe(de(()=>S),m(o=>({version:o.tag_name})),De({})),je(r).pipe(de(()=>S),m(o=>({stars:o.stargazers_count,forks:o.forks_count})),De({}))).pipe(m(([o,n])=>$($({},o),n)))}else{let r=`https://api.github.com/users/${e}`;return je(r).pipe(m(o=>({repositories:o.public_repos})),De({}))}}function bi(e,t){let r=`https://${e}/api/v4/projects/${encodeURIComponent(t)}`;return st(je(`${r}/releases/permalink/latest`).pipe(de(()=>S),m(({tag_name:o})=>({version:o})),De({})),je(r).pipe(de(()=>S),m(({star_count:o,forks_count:n})=>({stars:o,forks:n})),De({}))).pipe(m(([o,n])=>$($({},o),n)))}function vi(e){let t=e.match(/^.+github\.com\/([^/]+)\/?([^/]+)?/i);if(t){let[,r,o]=t;return hi(r,o)}if(t=e.match(/^.+?([^/]*gitlab[^/]+)\/(.+?)\/?$/i),t){let[,r,o]=t;return bi(r,o)}return S}var us;function ds(e){return us||(us=C(()=>{let t=__md_get("__source",sessionStorage);if(t)return I(t);if(ae("consent").length){let o=__md_get("__consent");if(!(o&&o.github))return S}return vi(e.href).pipe(w(o=>__md_set("__source",o,sessionStorage)))}).pipe(de(()=>S),b(t=>Object.keys(t).length>0),m(t=>({facts:t})),G(1)))}function gi(e){let t=R(":scope > :last-child",e);return C(()=>{let r=new g;return r.subscribe(({facts:o})=>{t.appendChild(_n(o)),t.classList.add("md-source__repository--active")}),ds(e).pipe(w(o=>r.next(o)),_(()=>r.complete()),m(o=>$({ref:e},o)))})}function hs(e,{viewport$:t,header$:r}){return ge(document.body).pipe(v(()=>mr(e,{header$:r,viewport$:t})),m(({offset:{y:o}})=>({hidden:o>=10})),ee("hidden"))}function yi(e,t){return C(()=>{let r=new g;return r.subscribe({next({hidden:o}){e.hidden=o},complete(){e.hidden=!1}}),(B("navigation.tabs.sticky")?I({hidden:!1}):hs(e,t)).pipe(w(o=>r.next(o)),_(()=>r.complete()),m(o=>$({ref:e},o)))})}function bs(e,{viewport$:t,header$:r}){let o=new Map,n=P(".md-nav__link",e);for(let s of n){let p=decodeURIComponent(s.hash.substring(1)),c=fe(`[id="${p}"]`);typeof c!="undefined"&&o.set(s,c)}let i=r.pipe(ee("height"),m(({height:s})=>{let p=Se("main"),c=R(":scope > :first-child",p);return s+.8*(c.offsetTop-p.offsetTop)}),pe());return ge(document.body).pipe(ee("height"),v(s=>C(()=>{let p=[];return I([...o].reduce((c,[l,f])=>{for(;p.length&&o.get(p[p.length-1]).tagName>=f.tagName;)p.pop();let u=f.offsetTop;for(;!u&&f.parentElement;)f=f.parentElement,u=f.offsetTop;let d=f.offsetParent;for(;d;d=d.offsetParent)u+=d.offsetTop;return c.set([...p=[...p,l]].reverse(),u)},new Map))}).pipe(m(p=>new Map([...p].sort(([,c],[,l])=>c-l))),He(i),v(([p,c])=>t.pipe(Fr(([l,f],{offset:{y:u},size:d})=>{let y=u+d.height>=Math.floor(s.height);for(;f.length;){let[,L]=f[0];if(L-c=u&&!y)f=[l.pop(),...f];else break}return[l,f]},[[],[...p]]),K((l,f)=>l[0]===f[0]&&l[1]===f[1])))))).pipe(m(([s,p])=>({prev:s.map(([c])=>c),next:p.map(([c])=>c)})),Q({prev:[],next:[]}),Be(2,1),m(([s,p])=>s.prev.length{let i=new g,a=i.pipe(Z(),ie(!0));if(i.subscribe(({prev:s,next:p})=>{for(let[c]of p)c.classList.remove("md-nav__link--passed"),c.classList.remove("md-nav__link--active");for(let[c,[l]]of s.entries())l.classList.add("md-nav__link--passed"),l.classList.toggle("md-nav__link--active",c===s.length-1)}),B("toc.follow")){let s=O(t.pipe(_e(1),m(()=>{})),t.pipe(_e(250),m(()=>"smooth")));i.pipe(b(({prev:p})=>p.length>0),He(o.pipe(ve(se))),re(s)).subscribe(([[{prev:p}],c])=>{let[l]=p[p.length-1];if(l.offsetHeight){let f=cr(l);if(typeof f!="undefined"){let u=l.offsetTop-f.offsetTop,{height:d}=ce(f);f.scrollTo({top:u-d/2,behavior:c})}}})}return B("navigation.tracking")&&t.pipe(W(a),ee("offset"),_e(250),Ce(1),W(n.pipe(Ce(1))),ct({delay:250}),re(i)).subscribe(([,{prev:s}])=>{let p=ye(),c=s[s.length-1];if(c&&c.length){let[l]=c,{hash:f}=new URL(l.href);p.hash!==f&&(p.hash=f,history.replaceState({},"",`${p}`))}else p.hash="",history.replaceState({},"",`${p}`)}),bs(e,{viewport$:t,header$:r}).pipe(w(s=>i.next(s)),_(()=>i.complete()),m(s=>$({ref:e},s)))})}function vs(e,{viewport$:t,main$:r,target$:o}){let n=t.pipe(m(({offset:{y:a}})=>a),Be(2,1),m(([a,s])=>a>s&&s>0),K()),i=r.pipe(m(({active:a})=>a));return z([i,n]).pipe(m(([a,s])=>!(a&&s)),K(),W(o.pipe(Ce(1))),ie(!0),ct({delay:250}),m(a=>({hidden:a})))}function Ei(e,{viewport$:t,header$:r,main$:o,target$:n}){let i=new g,a=i.pipe(Z(),ie(!0));return i.subscribe({next({hidden:s}){e.hidden=s,s?(e.setAttribute("tabindex","-1"),e.blur()):e.removeAttribute("tabindex")},complete(){e.style.top="",e.hidden=!0,e.removeAttribute("tabindex")}}),r.pipe(W(a),ee("height")).subscribe(({height:s})=>{e.style.top=`${s+16}px`}),h(e,"click").subscribe(s=>{s.preventDefault(),window.scrollTo({top:0})}),vs(e,{viewport$:t,main$:o,target$:n}).pipe(w(s=>i.next(s)),_(()=>i.complete()),m(s=>$({ref:e},s)))}function wi({document$:e,viewport$:t}){e.pipe(v(()=>P(".md-ellipsis")),ne(r=>tt(r).pipe(W(e.pipe(Ce(1))),b(o=>o),m(()=>r),Te(1))),b(r=>r.offsetWidth{let o=r.innerText,n=r.closest("a")||r;return n.title=o,B("content.tooltips")?mt(n,{viewport$:t}).pipe(W(e.pipe(Ce(1))),_(()=>n.removeAttribute("title"))):S})).subscribe(),B("content.tooltips")&&e.pipe(v(()=>P(".md-status")),ne(r=>mt(r,{viewport$:t}))).subscribe()}function Ti({document$:e,tablet$:t}){e.pipe(v(()=>P(".md-toggle--indeterminate")),w(r=>{r.indeterminate=!0,r.checked=!1}),ne(r=>h(r,"change").pipe(Dr(()=>r.classList.contains("md-toggle--indeterminate")),m(()=>r))),re(t)).subscribe(([r,o])=>{r.classList.remove("md-toggle--indeterminate"),o&&(r.checked=!1)})}function gs(){return/(iPad|iPhone|iPod)/.test(navigator.userAgent)}function Si({document$:e}){e.pipe(v(()=>P("[data-md-scrollfix]")),w(t=>t.removeAttribute("data-md-scrollfix")),b(gs),ne(t=>h(t,"touchstart").pipe(m(()=>t)))).subscribe(t=>{let r=t.scrollTop;r===0?t.scrollTop=1:r+t.offsetHeight===t.scrollHeight&&(t.scrollTop=r-1)})}function Oi({viewport$:e,tablet$:t}){z([ze("search"),t]).pipe(m(([r,o])=>r&&!o),v(r=>I(r).pipe(Ge(r?400:100))),re(e)).subscribe(([r,{offset:{y:o}}])=>{if(r)document.body.setAttribute("data-md-scrolllock",""),document.body.style.top=`-${o}px`;else{let n=-1*parseInt(document.body.style.top,10);document.body.removeAttribute("data-md-scrolllock"),document.body.style.top="",n&&window.scrollTo(0,n)}})}Object.entries||(Object.entries=function(e){let t=[];for(let r of Object.keys(e))t.push([r,e[r]]);return t});Object.values||(Object.values=function(e){let t=[];for(let r of Object.keys(e))t.push(e[r]);return t});typeof Element!="undefined"&&(Element.prototype.scrollTo||(Element.prototype.scrollTo=function(e,t){typeof e=="object"?(this.scrollLeft=e.left,this.scrollTop=e.top):(this.scrollLeft=e,this.scrollTop=t)}),Element.prototype.replaceWith||(Element.prototype.replaceWith=function(...e){let t=this.parentNode;if(t){e.length===0&&t.removeChild(this);for(let r=e.length-1;r>=0;r--){let o=e[r];typeof o=="string"?o=document.createTextNode(o):o.parentNode&&o.parentNode.removeChild(o),r?t.insertBefore(this.previousSibling,o):t.replaceChild(o,this)}}}));function ys(){return location.protocol==="file:"?Tt(`${new URL("search/search_index.js",eo.base)}`).pipe(m(()=>__index),G(1)):je(new URL("search/search_index.json",eo.base))}document.documentElement.classList.remove("no-js");document.documentElement.classList.add("js");var ot=Go(),Ut=sn(),Lt=ln(Ut),to=an(),Oe=gn(),hr=Pt("(min-width: 960px)"),Mi=Pt("(min-width: 1220px)"),_i=mn(),eo=xe(),Ai=document.forms.namedItem("search")?ys():Ye,ro=new g;Zn({alert$:ro});var oo=new g;B("navigation.instant")&&oi({location$:Ut,viewport$:Oe,progress$:oo}).subscribe(ot);var Li;((Li=eo.version)==null?void 0:Li.provider)==="mike"&&ci({document$:ot});O(Ut,Lt).pipe(Ge(125)).subscribe(()=>{Je("drawer",!1),Je("search",!1)});to.pipe(b(({mode:e})=>e==="global")).subscribe(e=>{switch(e.type){case"p":case",":let t=fe("link[rel=prev]");typeof t!="undefined"&<(t);break;case"n":case".":let r=fe("link[rel=next]");typeof r!="undefined"&<(r);break;case"Enter":let o=Ie();o instanceof HTMLLabelElement&&o.click()}});wi({viewport$:Oe,document$:ot});Ti({document$:ot,tablet$:hr});Si({document$:ot});Oi({viewport$:Oe,tablet$:hr});var rt=Kn(Se("header"),{viewport$:Oe}),Ft=ot.pipe(m(()=>Se("main")),v(e=>Gn(e,{viewport$:Oe,header$:rt})),G(1)),xs=O(...ae("consent").map(e=>En(e,{target$:Lt})),...ae("dialog").map(e=>qn(e,{alert$:ro})),...ae("palette").map(e=>Jn(e)),...ae("progress").map(e=>Xn(e,{progress$:oo})),...ae("search").map(e=>ui(e,{index$:Ai,keyboard$:to})),...ae("source").map(e=>gi(e))),Es=C(()=>O(...ae("announce").map(e=>xn(e)),...ae("content").map(e=>zn(e,{viewport$:Oe,target$:Lt,print$:_i})),...ae("content").map(e=>B("search.highlight")?di(e,{index$:Ai,location$:Ut}):S),...ae("header").map(e=>Yn(e,{viewport$:Oe,header$:rt,main$:Ft})),...ae("header-title").map(e=>Bn(e,{viewport$:Oe,header$:rt})),...ae("sidebar").map(e=>e.getAttribute("data-md-type")==="navigation"?Nr(Mi,()=>Zr(e,{viewport$:Oe,header$:rt,main$:Ft})):Nr(hr,()=>Zr(e,{viewport$:Oe,header$:rt,main$:Ft}))),...ae("tabs").map(e=>yi(e,{viewport$:Oe,header$:rt})),...ae("toc").map(e=>xi(e,{viewport$:Oe,header$:rt,main$:Ft,target$:Lt})),...ae("top").map(e=>Ei(e,{viewport$:Oe,header$:rt,main$:Ft,target$:Lt})))),Ci=ot.pipe(v(()=>Es),Re(xs),G(1));Ci.subscribe();window.document$=ot;window.location$=Ut;window.target$=Lt;window.keyboard$=to;window.viewport$=Oe;window.tablet$=hr;window.screen$=Mi;window.print$=_i;window.alert$=ro;window.progress$=oo;window.component$=Ci;})(); +//# sourceMappingURL=bundle.83f73b43.min.js.map + diff --git a/assets/javascripts/bundle.83f73b43.min.js.map b/assets/javascripts/bundle.83f73b43.min.js.map new file mode 100644 index 00000000..fe920b7d --- /dev/null +++ b/assets/javascripts/bundle.83f73b43.min.js.map @@ -0,0 +1,7 @@ +{ + "version": 3, + "sources": ["node_modules/focus-visible/dist/focus-visible.js", "node_modules/escape-html/index.js", "node_modules/clipboard/dist/clipboard.js", "src/templates/assets/javascripts/bundle.ts", "node_modules/tslib/tslib.es6.mjs", "node_modules/rxjs/src/internal/util/isFunction.ts", "node_modules/rxjs/src/internal/util/createErrorClass.ts", "node_modules/rxjs/src/internal/util/UnsubscriptionError.ts", "node_modules/rxjs/src/internal/util/arrRemove.ts", "node_modules/rxjs/src/internal/Subscription.ts", "node_modules/rxjs/src/internal/config.ts", "node_modules/rxjs/src/internal/scheduler/timeoutProvider.ts", "node_modules/rxjs/src/internal/util/reportUnhandledError.ts", "node_modules/rxjs/src/internal/util/noop.ts", "node_modules/rxjs/src/internal/NotificationFactories.ts", "node_modules/rxjs/src/internal/util/errorContext.ts", "node_modules/rxjs/src/internal/Subscriber.ts", "node_modules/rxjs/src/internal/symbol/observable.ts", "node_modules/rxjs/src/internal/util/identity.ts", "node_modules/rxjs/src/internal/util/pipe.ts", "node_modules/rxjs/src/internal/Observable.ts", "node_modules/rxjs/src/internal/util/lift.ts", "node_modules/rxjs/src/internal/operators/OperatorSubscriber.ts", "node_modules/rxjs/src/internal/scheduler/animationFrameProvider.ts", "node_modules/rxjs/src/internal/util/ObjectUnsubscribedError.ts", "node_modules/rxjs/src/internal/Subject.ts", "node_modules/rxjs/src/internal/BehaviorSubject.ts", "node_modules/rxjs/src/internal/scheduler/dateTimestampProvider.ts", "node_modules/rxjs/src/internal/ReplaySubject.ts", "node_modules/rxjs/src/internal/scheduler/Action.ts", "node_modules/rxjs/src/internal/scheduler/intervalProvider.ts", "node_modules/rxjs/src/internal/scheduler/AsyncAction.ts", "node_modules/rxjs/src/internal/Scheduler.ts", "node_modules/rxjs/src/internal/scheduler/AsyncScheduler.ts", "node_modules/rxjs/src/internal/scheduler/async.ts", "node_modules/rxjs/src/internal/scheduler/QueueAction.ts", "node_modules/rxjs/src/internal/scheduler/QueueScheduler.ts", "node_modules/rxjs/src/internal/scheduler/queue.ts", "node_modules/rxjs/src/internal/scheduler/AnimationFrameAction.ts", "node_modules/rxjs/src/internal/scheduler/AnimationFrameScheduler.ts", "node_modules/rxjs/src/internal/scheduler/animationFrame.ts", "node_modules/rxjs/src/internal/observable/empty.ts", "node_modules/rxjs/src/internal/util/isScheduler.ts", "node_modules/rxjs/src/internal/util/args.ts", "node_modules/rxjs/src/internal/util/isArrayLike.ts", "node_modules/rxjs/src/internal/util/isPromise.ts", "node_modules/rxjs/src/internal/util/isInteropObservable.ts", "node_modules/rxjs/src/internal/util/isAsyncIterable.ts", "node_modules/rxjs/src/internal/util/throwUnobservableError.ts", "node_modules/rxjs/src/internal/symbol/iterator.ts", "node_modules/rxjs/src/internal/util/isIterable.ts", "node_modules/rxjs/src/internal/util/isReadableStreamLike.ts", "node_modules/rxjs/src/internal/observable/innerFrom.ts", "node_modules/rxjs/src/internal/util/executeSchedule.ts", "node_modules/rxjs/src/internal/operators/observeOn.ts", "node_modules/rxjs/src/internal/operators/subscribeOn.ts", "node_modules/rxjs/src/internal/scheduled/scheduleObservable.ts", "node_modules/rxjs/src/internal/scheduled/schedulePromise.ts", "node_modules/rxjs/src/internal/scheduled/scheduleArray.ts", "node_modules/rxjs/src/internal/scheduled/scheduleIterable.ts", "node_modules/rxjs/src/internal/scheduled/scheduleAsyncIterable.ts", "node_modules/rxjs/src/internal/scheduled/scheduleReadableStreamLike.ts", "node_modules/rxjs/src/internal/scheduled/scheduled.ts", "node_modules/rxjs/src/internal/observable/from.ts", "node_modules/rxjs/src/internal/observable/of.ts", "node_modules/rxjs/src/internal/observable/throwError.ts", "node_modules/rxjs/src/internal/util/EmptyError.ts", "node_modules/rxjs/src/internal/util/isDate.ts", "node_modules/rxjs/src/internal/operators/map.ts", "node_modules/rxjs/src/internal/util/mapOneOrManyArgs.ts", "node_modules/rxjs/src/internal/util/argsArgArrayOrObject.ts", "node_modules/rxjs/src/internal/util/createObject.ts", "node_modules/rxjs/src/internal/observable/combineLatest.ts", "node_modules/rxjs/src/internal/operators/mergeInternals.ts", "node_modules/rxjs/src/internal/operators/mergeMap.ts", "node_modules/rxjs/src/internal/operators/mergeAll.ts", "node_modules/rxjs/src/internal/operators/concatAll.ts", "node_modules/rxjs/src/internal/observable/concat.ts", "node_modules/rxjs/src/internal/observable/defer.ts", "node_modules/rxjs/src/internal/observable/fromEvent.ts", "node_modules/rxjs/src/internal/observable/fromEventPattern.ts", "node_modules/rxjs/src/internal/observable/timer.ts", "node_modules/rxjs/src/internal/observable/merge.ts", "node_modules/rxjs/src/internal/observable/never.ts", "node_modules/rxjs/src/internal/util/argsOrArgArray.ts", "node_modules/rxjs/src/internal/operators/filter.ts", "node_modules/rxjs/src/internal/observable/zip.ts", "node_modules/rxjs/src/internal/operators/audit.ts", "node_modules/rxjs/src/internal/operators/auditTime.ts", "node_modules/rxjs/src/internal/operators/bufferCount.ts", "node_modules/rxjs/src/internal/operators/catchError.ts", "node_modules/rxjs/src/internal/operators/scanInternals.ts", "node_modules/rxjs/src/internal/operators/combineLatest.ts", "node_modules/rxjs/src/internal/operators/combineLatestWith.ts", "node_modules/rxjs/src/internal/operators/debounce.ts", "node_modules/rxjs/src/internal/operators/debounceTime.ts", "node_modules/rxjs/src/internal/operators/defaultIfEmpty.ts", "node_modules/rxjs/src/internal/operators/take.ts", "node_modules/rxjs/src/internal/operators/ignoreElements.ts", "node_modules/rxjs/src/internal/operators/mapTo.ts", "node_modules/rxjs/src/internal/operators/delayWhen.ts", "node_modules/rxjs/src/internal/operators/delay.ts", "node_modules/rxjs/src/internal/operators/distinctUntilChanged.ts", "node_modules/rxjs/src/internal/operators/distinctUntilKeyChanged.ts", "node_modules/rxjs/src/internal/operators/throwIfEmpty.ts", "node_modules/rxjs/src/internal/operators/endWith.ts", "node_modules/rxjs/src/internal/operators/finalize.ts", "node_modules/rxjs/src/internal/operators/first.ts", "node_modules/rxjs/src/internal/operators/takeLast.ts", "node_modules/rxjs/src/internal/operators/merge.ts", "node_modules/rxjs/src/internal/operators/mergeWith.ts", "node_modules/rxjs/src/internal/operators/repeat.ts", "node_modules/rxjs/src/internal/operators/scan.ts", "node_modules/rxjs/src/internal/operators/share.ts", "node_modules/rxjs/src/internal/operators/shareReplay.ts", "node_modules/rxjs/src/internal/operators/skip.ts", "node_modules/rxjs/src/internal/operators/skipUntil.ts", "node_modules/rxjs/src/internal/operators/startWith.ts", "node_modules/rxjs/src/internal/operators/switchMap.ts", "node_modules/rxjs/src/internal/operators/takeUntil.ts", "node_modules/rxjs/src/internal/operators/takeWhile.ts", "node_modules/rxjs/src/internal/operators/tap.ts", "node_modules/rxjs/src/internal/operators/throttle.ts", "node_modules/rxjs/src/internal/operators/throttleTime.ts", "node_modules/rxjs/src/internal/operators/withLatestFrom.ts", "node_modules/rxjs/src/internal/operators/zip.ts", "node_modules/rxjs/src/internal/operators/zipWith.ts", "src/templates/assets/javascripts/browser/document/index.ts", "src/templates/assets/javascripts/browser/element/_/index.ts", "src/templates/assets/javascripts/browser/element/focus/index.ts", "src/templates/assets/javascripts/browser/element/hover/index.ts", "src/templates/assets/javascripts/utilities/h/index.ts", "src/templates/assets/javascripts/utilities/round/index.ts", "src/templates/assets/javascripts/browser/script/index.ts", "src/templates/assets/javascripts/browser/element/size/_/index.ts", "src/templates/assets/javascripts/browser/element/size/content/index.ts", "src/templates/assets/javascripts/browser/element/offset/_/index.ts", "src/templates/assets/javascripts/browser/element/offset/content/index.ts", "src/templates/assets/javascripts/browser/element/visibility/index.ts", "src/templates/assets/javascripts/browser/toggle/index.ts", "src/templates/assets/javascripts/browser/keyboard/index.ts", "src/templates/assets/javascripts/browser/location/_/index.ts", "src/templates/assets/javascripts/browser/location/hash/index.ts", "src/templates/assets/javascripts/browser/media/index.ts", "src/templates/assets/javascripts/browser/request/index.ts", "src/templates/assets/javascripts/browser/viewport/offset/index.ts", "src/templates/assets/javascripts/browser/viewport/size/index.ts", "src/templates/assets/javascripts/browser/viewport/_/index.ts", "src/templates/assets/javascripts/browser/viewport/at/index.ts", "src/templates/assets/javascripts/browser/worker/index.ts", "src/templates/assets/javascripts/_/index.ts", "src/templates/assets/javascripts/components/_/index.ts", "src/templates/assets/javascripts/components/announce/index.ts", "src/templates/assets/javascripts/components/consent/index.ts", "src/templates/assets/javascripts/templates/tooltip/index.tsx", "src/templates/assets/javascripts/templates/annotation/index.tsx", "src/templates/assets/javascripts/templates/clipboard/index.tsx", "src/templates/assets/javascripts/templates/search/index.tsx", "src/templates/assets/javascripts/templates/source/index.tsx", "src/templates/assets/javascripts/templates/tabbed/index.tsx", "src/templates/assets/javascripts/templates/table/index.tsx", "src/templates/assets/javascripts/templates/version/index.tsx", "src/templates/assets/javascripts/components/tooltip2/index.ts", "src/templates/assets/javascripts/components/content/annotation/_/index.ts", "src/templates/assets/javascripts/components/content/annotation/list/index.ts", "src/templates/assets/javascripts/components/content/annotation/block/index.ts", "src/templates/assets/javascripts/components/content/code/_/index.ts", "src/templates/assets/javascripts/components/content/details/index.ts", "src/templates/assets/javascripts/components/content/mermaid/index.css", "src/templates/assets/javascripts/components/content/mermaid/index.ts", "src/templates/assets/javascripts/components/content/table/index.ts", "src/templates/assets/javascripts/components/content/tabs/index.ts", "src/templates/assets/javascripts/components/content/_/index.ts", "src/templates/assets/javascripts/components/dialog/index.ts", "src/templates/assets/javascripts/components/tooltip/index.ts", "src/templates/assets/javascripts/components/header/_/index.ts", "src/templates/assets/javascripts/components/header/title/index.ts", "src/templates/assets/javascripts/components/main/index.ts", "src/templates/assets/javascripts/components/palette/index.ts", "src/templates/assets/javascripts/components/progress/index.ts", "src/templates/assets/javascripts/integrations/clipboard/index.ts", "src/templates/assets/javascripts/integrations/sitemap/index.ts", "src/templates/assets/javascripts/integrations/instant/index.ts", "src/templates/assets/javascripts/integrations/search/highlighter/index.ts", "src/templates/assets/javascripts/integrations/search/worker/message/index.ts", "src/templates/assets/javascripts/integrations/search/worker/_/index.ts", "src/templates/assets/javascripts/integrations/version/findurl/index.ts", "src/templates/assets/javascripts/integrations/version/index.ts", "src/templates/assets/javascripts/components/search/query/index.ts", "src/templates/assets/javascripts/components/search/result/index.ts", "src/templates/assets/javascripts/components/search/share/index.ts", "src/templates/assets/javascripts/components/search/suggest/index.ts", "src/templates/assets/javascripts/components/search/_/index.ts", "src/templates/assets/javascripts/components/search/highlight/index.ts", "src/templates/assets/javascripts/components/sidebar/index.ts", "src/templates/assets/javascripts/components/source/facts/github/index.ts", "src/templates/assets/javascripts/components/source/facts/gitlab/index.ts", "src/templates/assets/javascripts/components/source/facts/_/index.ts", "src/templates/assets/javascripts/components/source/_/index.ts", "src/templates/assets/javascripts/components/tabs/index.ts", "src/templates/assets/javascripts/components/toc/index.ts", "src/templates/assets/javascripts/components/top/index.ts", "src/templates/assets/javascripts/patches/ellipsis/index.ts", "src/templates/assets/javascripts/patches/indeterminate/index.ts", "src/templates/assets/javascripts/patches/scrollfix/index.ts", "src/templates/assets/javascripts/patches/scrolllock/index.ts", "src/templates/assets/javascripts/polyfills/index.ts"], + "sourcesContent": ["(function (global, factory) {\n typeof exports === 'object' && typeof module !== 'undefined' ? factory() :\n typeof define === 'function' && define.amd ? define(factory) :\n (factory());\n}(this, (function () { 'use strict';\n\n /**\n * Applies the :focus-visible polyfill at the given scope.\n * A scope in this case is either the top-level Document or a Shadow Root.\n *\n * @param {(Document|ShadowRoot)} scope\n * @see https://github.com/WICG/focus-visible\n */\n function applyFocusVisiblePolyfill(scope) {\n var hadKeyboardEvent = true;\n var hadFocusVisibleRecently = false;\n var hadFocusVisibleRecentlyTimeout = null;\n\n var inputTypesAllowlist = {\n text: true,\n search: true,\n url: true,\n tel: true,\n email: true,\n password: true,\n number: true,\n date: true,\n month: true,\n week: true,\n time: true,\n datetime: true,\n 'datetime-local': true\n };\n\n /**\n * Helper function for legacy browsers and iframes which sometimes focus\n * elements like document, body, and non-interactive SVG.\n * @param {Element} el\n */\n function isValidFocusTarget(el) {\n if (\n el &&\n el !== document &&\n el.nodeName !== 'HTML' &&\n el.nodeName !== 'BODY' &&\n 'classList' in el &&\n 'contains' in el.classList\n ) {\n return true;\n }\n return false;\n }\n\n /**\n * Computes whether the given element should automatically trigger the\n * `focus-visible` class being added, i.e. whether it should always match\n * `:focus-visible` when focused.\n * @param {Element} el\n * @return {boolean}\n */\n function focusTriggersKeyboardModality(el) {\n var type = el.type;\n var tagName = el.tagName;\n\n if (tagName === 'INPUT' && inputTypesAllowlist[type] && !el.readOnly) {\n return true;\n }\n\n if (tagName === 'TEXTAREA' && !el.readOnly) {\n return true;\n }\n\n if (el.isContentEditable) {\n return true;\n }\n\n return false;\n }\n\n /**\n * Add the `focus-visible` class to the given element if it was not added by\n * the author.\n * @param {Element} el\n */\n function addFocusVisibleClass(el) {\n if (el.classList.contains('focus-visible')) {\n return;\n }\n el.classList.add('focus-visible');\n el.setAttribute('data-focus-visible-added', '');\n }\n\n /**\n * Remove the `focus-visible` class from the given element if it was not\n * originally added by the author.\n * @param {Element} el\n */\n function removeFocusVisibleClass(el) {\n if (!el.hasAttribute('data-focus-visible-added')) {\n return;\n }\n el.classList.remove('focus-visible');\n el.removeAttribute('data-focus-visible-added');\n }\n\n /**\n * If the most recent user interaction was via the keyboard;\n * and the key press did not include a meta, alt/option, or control key;\n * then the modality is keyboard. Otherwise, the modality is not keyboard.\n * Apply `focus-visible` to any current active element and keep track\n * of our keyboard modality state with `hadKeyboardEvent`.\n * @param {KeyboardEvent} e\n */\n function onKeyDown(e) {\n if (e.metaKey || e.altKey || e.ctrlKey) {\n return;\n }\n\n if (isValidFocusTarget(scope.activeElement)) {\n addFocusVisibleClass(scope.activeElement);\n }\n\n hadKeyboardEvent = true;\n }\n\n /**\n * If at any point a user clicks with a pointing device, ensure that we change\n * the modality away from keyboard.\n * This avoids the situation where a user presses a key on an already focused\n * element, and then clicks on a different element, focusing it with a\n * pointing device, while we still think we're in keyboard modality.\n * @param {Event} e\n */\n function onPointerDown(e) {\n hadKeyboardEvent = false;\n }\n\n /**\n * On `focus`, add the `focus-visible` class to the target if:\n * - the target received focus as a result of keyboard navigation, or\n * - the event target is an element that will likely require interaction\n * via the keyboard (e.g. a text box)\n * @param {Event} e\n */\n function onFocus(e) {\n // Prevent IE from focusing the document or HTML element.\n if (!isValidFocusTarget(e.target)) {\n return;\n }\n\n if (hadKeyboardEvent || focusTriggersKeyboardModality(e.target)) {\n addFocusVisibleClass(e.target);\n }\n }\n\n /**\n * On `blur`, remove the `focus-visible` class from the target.\n * @param {Event} e\n */\n function onBlur(e) {\n if (!isValidFocusTarget(e.target)) {\n return;\n }\n\n if (\n e.target.classList.contains('focus-visible') ||\n e.target.hasAttribute('data-focus-visible-added')\n ) {\n // To detect a tab/window switch, we look for a blur event followed\n // rapidly by a visibility change.\n // If we don't see a visibility change within 100ms, it's probably a\n // regular focus change.\n hadFocusVisibleRecently = true;\n window.clearTimeout(hadFocusVisibleRecentlyTimeout);\n hadFocusVisibleRecentlyTimeout = window.setTimeout(function() {\n hadFocusVisibleRecently = false;\n }, 100);\n removeFocusVisibleClass(e.target);\n }\n }\n\n /**\n * If the user changes tabs, keep track of whether or not the previously\n * focused element had .focus-visible.\n * @param {Event} e\n */\n function onVisibilityChange(e) {\n if (document.visibilityState === 'hidden') {\n // If the tab becomes active again, the browser will handle calling focus\n // on the element (Safari actually calls it twice).\n // If this tab change caused a blur on an element with focus-visible,\n // re-apply the class when the user switches back to the tab.\n if (hadFocusVisibleRecently) {\n hadKeyboardEvent = true;\n }\n addInitialPointerMoveListeners();\n }\n }\n\n /**\n * Add a group of listeners to detect usage of any pointing devices.\n * These listeners will be added when the polyfill first loads, and anytime\n * the window is blurred, so that they are active when the window regains\n * focus.\n */\n function addInitialPointerMoveListeners() {\n document.addEventListener('mousemove', onInitialPointerMove);\n document.addEventListener('mousedown', onInitialPointerMove);\n document.addEventListener('mouseup', onInitialPointerMove);\n document.addEventListener('pointermove', onInitialPointerMove);\n document.addEventListener('pointerdown', onInitialPointerMove);\n document.addEventListener('pointerup', onInitialPointerMove);\n document.addEventListener('touchmove', onInitialPointerMove);\n document.addEventListener('touchstart', onInitialPointerMove);\n document.addEventListener('touchend', onInitialPointerMove);\n }\n\n function removeInitialPointerMoveListeners() {\n document.removeEventListener('mousemove', onInitialPointerMove);\n document.removeEventListener('mousedown', onInitialPointerMove);\n document.removeEventListener('mouseup', onInitialPointerMove);\n document.removeEventListener('pointermove', onInitialPointerMove);\n document.removeEventListener('pointerdown', onInitialPointerMove);\n document.removeEventListener('pointerup', onInitialPointerMove);\n document.removeEventListener('touchmove', onInitialPointerMove);\n document.removeEventListener('touchstart', onInitialPointerMove);\n document.removeEventListener('touchend', onInitialPointerMove);\n }\n\n /**\n * When the polfyill first loads, assume the user is in keyboard modality.\n * If any event is received from a pointing device (e.g. mouse, pointer,\n * touch), turn off keyboard modality.\n * This accounts for situations where focus enters the page from the URL bar.\n * @param {Event} e\n */\n function onInitialPointerMove(e) {\n // Work around a Safari quirk that fires a mousemove on whenever the\n // window blurs, even if you're tabbing out of the page. \u00AF\\_(\u30C4)_/\u00AF\n if (e.target.nodeName && e.target.nodeName.toLowerCase() === 'html') {\n return;\n }\n\n hadKeyboardEvent = false;\n removeInitialPointerMoveListeners();\n }\n\n // For some kinds of state, we are interested in changes at the global scope\n // only. For example, global pointer input, global key presses and global\n // visibility change should affect the state at every scope:\n document.addEventListener('keydown', onKeyDown, true);\n document.addEventListener('mousedown', onPointerDown, true);\n document.addEventListener('pointerdown', onPointerDown, true);\n document.addEventListener('touchstart', onPointerDown, true);\n document.addEventListener('visibilitychange', onVisibilityChange, true);\n\n addInitialPointerMoveListeners();\n\n // For focus and blur, we specifically care about state changes in the local\n // scope. This is because focus / blur events that originate from within a\n // shadow root are not re-dispatched from the host element if it was already\n // the active element in its own scope:\n scope.addEventListener('focus', onFocus, true);\n scope.addEventListener('blur', onBlur, true);\n\n // We detect that a node is a ShadowRoot by ensuring that it is a\n // DocumentFragment and also has a host property. This check covers native\n // implementation and polyfill implementation transparently. If we only cared\n // about the native implementation, we could just check if the scope was\n // an instance of a ShadowRoot.\n if (scope.nodeType === Node.DOCUMENT_FRAGMENT_NODE && scope.host) {\n // Since a ShadowRoot is a special kind of DocumentFragment, it does not\n // have a root element to add a class to. So, we add this attribute to the\n // host element instead:\n scope.host.setAttribute('data-js-focus-visible', '');\n } else if (scope.nodeType === Node.DOCUMENT_NODE) {\n document.documentElement.classList.add('js-focus-visible');\n document.documentElement.setAttribute('data-js-focus-visible', '');\n }\n }\n\n // It is important to wrap all references to global window and document in\n // these checks to support server-side rendering use cases\n // @see https://github.com/WICG/focus-visible/issues/199\n if (typeof window !== 'undefined' && typeof document !== 'undefined') {\n // Make the polyfill helper globally available. This can be used as a signal\n // to interested libraries that wish to coordinate with the polyfill for e.g.,\n // applying the polyfill to a shadow root:\n window.applyFocusVisiblePolyfill = applyFocusVisiblePolyfill;\n\n // Notify interested libraries of the polyfill's presence, in case the\n // polyfill was loaded lazily:\n var event;\n\n try {\n event = new CustomEvent('focus-visible-polyfill-ready');\n } catch (error) {\n // IE11 does not support using CustomEvent as a constructor directly:\n event = document.createEvent('CustomEvent');\n event.initCustomEvent('focus-visible-polyfill-ready', false, false, {});\n }\n\n window.dispatchEvent(event);\n }\n\n if (typeof document !== 'undefined') {\n // Apply the polyfill to the global document, so that no JavaScript\n // coordination is required to use the polyfill in the top-level document:\n applyFocusVisiblePolyfill(document);\n }\n\n})));\n", "/*!\n * escape-html\n * Copyright(c) 2012-2013 TJ Holowaychuk\n * Copyright(c) 2015 Andreas Lubbe\n * Copyright(c) 2015 Tiancheng \"Timothy\" Gu\n * MIT Licensed\n */\n\n'use strict';\n\n/**\n * Module variables.\n * @private\n */\n\nvar matchHtmlRegExp = /[\"'&<>]/;\n\n/**\n * Module exports.\n * @public\n */\n\nmodule.exports = escapeHtml;\n\n/**\n * Escape special characters in the given string of html.\n *\n * @param {string} string The string to escape for inserting into HTML\n * @return {string}\n * @public\n */\n\nfunction escapeHtml(string) {\n var str = '' + string;\n var match = matchHtmlRegExp.exec(str);\n\n if (!match) {\n return str;\n }\n\n var escape;\n var html = '';\n var index = 0;\n var lastIndex = 0;\n\n for (index = match.index; index < str.length; index++) {\n switch (str.charCodeAt(index)) {\n case 34: // \"\n escape = '"';\n break;\n case 38: // &\n escape = '&';\n break;\n case 39: // '\n escape = ''';\n break;\n case 60: // <\n escape = '<';\n break;\n case 62: // >\n escape = '>';\n break;\n default:\n continue;\n }\n\n if (lastIndex !== index) {\n html += str.substring(lastIndex, index);\n }\n\n lastIndex = index + 1;\n html += escape;\n }\n\n return lastIndex !== index\n ? html + str.substring(lastIndex, index)\n : html;\n}\n", "/*!\n * clipboard.js v2.0.11\n * https://clipboardjs.com/\n *\n * Licensed MIT \u00A9 Zeno Rocha\n */\n(function webpackUniversalModuleDefinition(root, factory) {\n\tif(typeof exports === 'object' && typeof module === 'object')\n\t\tmodule.exports = factory();\n\telse if(typeof define === 'function' && define.amd)\n\t\tdefine([], factory);\n\telse if(typeof exports === 'object')\n\t\texports[\"ClipboardJS\"] = factory();\n\telse\n\t\troot[\"ClipboardJS\"] = factory();\n})(this, function() {\nreturn /******/ (function() { // webpackBootstrap\n/******/ \tvar __webpack_modules__ = ({\n\n/***/ 686:\n/***/ (function(__unused_webpack_module, __webpack_exports__, __webpack_require__) {\n\n\"use strict\";\n\n// EXPORTS\n__webpack_require__.d(__webpack_exports__, {\n \"default\": function() { return /* binding */ clipboard; }\n});\n\n// EXTERNAL MODULE: ./node_modules/tiny-emitter/index.js\nvar tiny_emitter = __webpack_require__(279);\nvar tiny_emitter_default = /*#__PURE__*/__webpack_require__.n(tiny_emitter);\n// EXTERNAL MODULE: ./node_modules/good-listener/src/listen.js\nvar listen = __webpack_require__(370);\nvar listen_default = /*#__PURE__*/__webpack_require__.n(listen);\n// EXTERNAL MODULE: ./node_modules/select/src/select.js\nvar src_select = __webpack_require__(817);\nvar select_default = /*#__PURE__*/__webpack_require__.n(src_select);\n;// CONCATENATED MODULE: ./src/common/command.js\n/**\n * Executes a given operation type.\n * @param {String} type\n * @return {Boolean}\n */\nfunction command(type) {\n try {\n return document.execCommand(type);\n } catch (err) {\n return false;\n }\n}\n;// CONCATENATED MODULE: ./src/actions/cut.js\n\n\n/**\n * Cut action wrapper.\n * @param {String|HTMLElement} target\n * @return {String}\n */\n\nvar ClipboardActionCut = function ClipboardActionCut(target) {\n var selectedText = select_default()(target);\n command('cut');\n return selectedText;\n};\n\n/* harmony default export */ var actions_cut = (ClipboardActionCut);\n;// CONCATENATED MODULE: ./src/common/create-fake-element.js\n/**\n * Creates a fake textarea element with a value.\n * @param {String} value\n * @return {HTMLElement}\n */\nfunction createFakeElement(value) {\n var isRTL = document.documentElement.getAttribute('dir') === 'rtl';\n var fakeElement = document.createElement('textarea'); // Prevent zooming on iOS\n\n fakeElement.style.fontSize = '12pt'; // Reset box model\n\n fakeElement.style.border = '0';\n fakeElement.style.padding = '0';\n fakeElement.style.margin = '0'; // Move element out of screen horizontally\n\n fakeElement.style.position = 'absolute';\n fakeElement.style[isRTL ? 'right' : 'left'] = '-9999px'; // Move element to the same position vertically\n\n var yPosition = window.pageYOffset || document.documentElement.scrollTop;\n fakeElement.style.top = \"\".concat(yPosition, \"px\");\n fakeElement.setAttribute('readonly', '');\n fakeElement.value = value;\n return fakeElement;\n}\n;// CONCATENATED MODULE: ./src/actions/copy.js\n\n\n\n/**\n * Create fake copy action wrapper using a fake element.\n * @param {String} target\n * @param {Object} options\n * @return {String}\n */\n\nvar fakeCopyAction = function fakeCopyAction(value, options) {\n var fakeElement = createFakeElement(value);\n options.container.appendChild(fakeElement);\n var selectedText = select_default()(fakeElement);\n command('copy');\n fakeElement.remove();\n return selectedText;\n};\n/**\n * Copy action wrapper.\n * @param {String|HTMLElement} target\n * @param {Object} options\n * @return {String}\n */\n\n\nvar ClipboardActionCopy = function ClipboardActionCopy(target) {\n var options = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : {\n container: document.body\n };\n var selectedText = '';\n\n if (typeof target === 'string') {\n selectedText = fakeCopyAction(target, options);\n } else if (target instanceof HTMLInputElement && !['text', 'search', 'url', 'tel', 'password'].includes(target === null || target === void 0 ? void 0 : target.type)) {\n // If input type doesn't support `setSelectionRange`. Simulate it. https://developer.mozilla.org/en-US/docs/Web/API/HTMLInputElement/setSelectionRange\n selectedText = fakeCopyAction(target.value, options);\n } else {\n selectedText = select_default()(target);\n command('copy');\n }\n\n return selectedText;\n};\n\n/* harmony default export */ var actions_copy = (ClipboardActionCopy);\n;// CONCATENATED MODULE: ./src/actions/default.js\nfunction _typeof(obj) { \"@babel/helpers - typeof\"; if (typeof Symbol === \"function\" && typeof Symbol.iterator === \"symbol\") { _typeof = function _typeof(obj) { return typeof obj; }; } else { _typeof = function _typeof(obj) { return obj && typeof Symbol === \"function\" && obj.constructor === Symbol && obj !== Symbol.prototype ? \"symbol\" : typeof obj; }; } return _typeof(obj); }\n\n\n\n/**\n * Inner function which performs selection from either `text` or `target`\n * properties and then executes copy or cut operations.\n * @param {Object} options\n */\n\nvar ClipboardActionDefault = function ClipboardActionDefault() {\n var options = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : {};\n // Defines base properties passed from constructor.\n var _options$action = options.action,\n action = _options$action === void 0 ? 'copy' : _options$action,\n container = options.container,\n target = options.target,\n text = options.text; // Sets the `action` to be performed which can be either 'copy' or 'cut'.\n\n if (action !== 'copy' && action !== 'cut') {\n throw new Error('Invalid \"action\" value, use either \"copy\" or \"cut\"');\n } // Sets the `target` property using an element that will be have its content copied.\n\n\n if (target !== undefined) {\n if (target && _typeof(target) === 'object' && target.nodeType === 1) {\n if (action === 'copy' && target.hasAttribute('disabled')) {\n throw new Error('Invalid \"target\" attribute. Please use \"readonly\" instead of \"disabled\" attribute');\n }\n\n if (action === 'cut' && (target.hasAttribute('readonly') || target.hasAttribute('disabled'))) {\n throw new Error('Invalid \"target\" attribute. You can\\'t cut text from elements with \"readonly\" or \"disabled\" attributes');\n }\n } else {\n throw new Error('Invalid \"target\" value, use a valid Element');\n }\n } // Define selection strategy based on `text` property.\n\n\n if (text) {\n return actions_copy(text, {\n container: container\n });\n } // Defines which selection strategy based on `target` property.\n\n\n if (target) {\n return action === 'cut' ? actions_cut(target) : actions_copy(target, {\n container: container\n });\n }\n};\n\n/* harmony default export */ var actions_default = (ClipboardActionDefault);\n;// CONCATENATED MODULE: ./src/clipboard.js\nfunction clipboard_typeof(obj) { \"@babel/helpers - typeof\"; if (typeof Symbol === \"function\" && typeof Symbol.iterator === \"symbol\") { clipboard_typeof = function _typeof(obj) { return typeof obj; }; } else { clipboard_typeof = function _typeof(obj) { return obj && typeof Symbol === \"function\" && obj.constructor === Symbol && obj !== Symbol.prototype ? \"symbol\" : typeof obj; }; } return clipboard_typeof(obj); }\n\nfunction _classCallCheck(instance, Constructor) { if (!(instance instanceof Constructor)) { throw new TypeError(\"Cannot call a class as a function\"); } }\n\nfunction _defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if (\"value\" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } }\n\nfunction _createClass(Constructor, protoProps, staticProps) { if (protoProps) _defineProperties(Constructor.prototype, protoProps); if (staticProps) _defineProperties(Constructor, staticProps); return Constructor; }\n\nfunction _inherits(subClass, superClass) { if (typeof superClass !== \"function\" && superClass !== null) { throw new TypeError(\"Super expression must either be null or a function\"); } subClass.prototype = Object.create(superClass && superClass.prototype, { constructor: { value: subClass, writable: true, configurable: true } }); if (superClass) _setPrototypeOf(subClass, superClass); }\n\nfunction _setPrototypeOf(o, p) { _setPrototypeOf = Object.setPrototypeOf || function _setPrototypeOf(o, p) { o.__proto__ = p; return o; }; return _setPrototypeOf(o, p); }\n\nfunction _createSuper(Derived) { var hasNativeReflectConstruct = _isNativeReflectConstruct(); return function _createSuperInternal() { var Super = _getPrototypeOf(Derived), result; if (hasNativeReflectConstruct) { var NewTarget = _getPrototypeOf(this).constructor; result = Reflect.construct(Super, arguments, NewTarget); } else { result = Super.apply(this, arguments); } return _possibleConstructorReturn(this, result); }; }\n\nfunction _possibleConstructorReturn(self, call) { if (call && (clipboard_typeof(call) === \"object\" || typeof call === \"function\")) { return call; } return _assertThisInitialized(self); }\n\nfunction _assertThisInitialized(self) { if (self === void 0) { throw new ReferenceError(\"this hasn't been initialised - super() hasn't been called\"); } return self; }\n\nfunction _isNativeReflectConstruct() { if (typeof Reflect === \"undefined\" || !Reflect.construct) return false; if (Reflect.construct.sham) return false; if (typeof Proxy === \"function\") return true; try { Date.prototype.toString.call(Reflect.construct(Date, [], function () {})); return true; } catch (e) { return false; } }\n\nfunction _getPrototypeOf(o) { _getPrototypeOf = Object.setPrototypeOf ? Object.getPrototypeOf : function _getPrototypeOf(o) { return o.__proto__ || Object.getPrototypeOf(o); }; return _getPrototypeOf(o); }\n\n\n\n\n\n\n/**\n * Helper function to retrieve attribute value.\n * @param {String} suffix\n * @param {Element} element\n */\n\nfunction getAttributeValue(suffix, element) {\n var attribute = \"data-clipboard-\".concat(suffix);\n\n if (!element.hasAttribute(attribute)) {\n return;\n }\n\n return element.getAttribute(attribute);\n}\n/**\n * Base class which takes one or more elements, adds event listeners to them,\n * and instantiates a new `ClipboardAction` on each click.\n */\n\n\nvar Clipboard = /*#__PURE__*/function (_Emitter) {\n _inherits(Clipboard, _Emitter);\n\n var _super = _createSuper(Clipboard);\n\n /**\n * @param {String|HTMLElement|HTMLCollection|NodeList} trigger\n * @param {Object} options\n */\n function Clipboard(trigger, options) {\n var _this;\n\n _classCallCheck(this, Clipboard);\n\n _this = _super.call(this);\n\n _this.resolveOptions(options);\n\n _this.listenClick(trigger);\n\n return _this;\n }\n /**\n * Defines if attributes would be resolved using internal setter functions\n * or custom functions that were passed in the constructor.\n * @param {Object} options\n */\n\n\n _createClass(Clipboard, [{\n key: \"resolveOptions\",\n value: function resolveOptions() {\n var options = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : {};\n this.action = typeof options.action === 'function' ? options.action : this.defaultAction;\n this.target = typeof options.target === 'function' ? options.target : this.defaultTarget;\n this.text = typeof options.text === 'function' ? options.text : this.defaultText;\n this.container = clipboard_typeof(options.container) === 'object' ? options.container : document.body;\n }\n /**\n * Adds a click event listener to the passed trigger.\n * @param {String|HTMLElement|HTMLCollection|NodeList} trigger\n */\n\n }, {\n key: \"listenClick\",\n value: function listenClick(trigger) {\n var _this2 = this;\n\n this.listener = listen_default()(trigger, 'click', function (e) {\n return _this2.onClick(e);\n });\n }\n /**\n * Defines a new `ClipboardAction` on each click event.\n * @param {Event} e\n */\n\n }, {\n key: \"onClick\",\n value: function onClick(e) {\n var trigger = e.delegateTarget || e.currentTarget;\n var action = this.action(trigger) || 'copy';\n var text = actions_default({\n action: action,\n container: this.container,\n target: this.target(trigger),\n text: this.text(trigger)\n }); // Fires an event based on the copy operation result.\n\n this.emit(text ? 'success' : 'error', {\n action: action,\n text: text,\n trigger: trigger,\n clearSelection: function clearSelection() {\n if (trigger) {\n trigger.focus();\n }\n\n window.getSelection().removeAllRanges();\n }\n });\n }\n /**\n * Default `action` lookup function.\n * @param {Element} trigger\n */\n\n }, {\n key: \"defaultAction\",\n value: function defaultAction(trigger) {\n return getAttributeValue('action', trigger);\n }\n /**\n * Default `target` lookup function.\n * @param {Element} trigger\n */\n\n }, {\n key: \"defaultTarget\",\n value: function defaultTarget(trigger) {\n var selector = getAttributeValue('target', trigger);\n\n if (selector) {\n return document.querySelector(selector);\n }\n }\n /**\n * Allow fire programmatically a copy action\n * @param {String|HTMLElement} target\n * @param {Object} options\n * @returns Text copied.\n */\n\n }, {\n key: \"defaultText\",\n\n /**\n * Default `text` lookup function.\n * @param {Element} trigger\n */\n value: function defaultText(trigger) {\n return getAttributeValue('text', trigger);\n }\n /**\n * Destroy lifecycle.\n */\n\n }, {\n key: \"destroy\",\n value: function destroy() {\n this.listener.destroy();\n }\n }], [{\n key: \"copy\",\n value: function copy(target) {\n var options = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : {\n container: document.body\n };\n return actions_copy(target, options);\n }\n /**\n * Allow fire programmatically a cut action\n * @param {String|HTMLElement} target\n * @returns Text cutted.\n */\n\n }, {\n key: \"cut\",\n value: function cut(target) {\n return actions_cut(target);\n }\n /**\n * Returns the support of the given action, or all actions if no action is\n * given.\n * @param {String} [action]\n */\n\n }, {\n key: \"isSupported\",\n value: function isSupported() {\n var action = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : ['copy', 'cut'];\n var actions = typeof action === 'string' ? [action] : action;\n var support = !!document.queryCommandSupported;\n actions.forEach(function (action) {\n support = support && !!document.queryCommandSupported(action);\n });\n return support;\n }\n }]);\n\n return Clipboard;\n}((tiny_emitter_default()));\n\n/* harmony default export */ var clipboard = (Clipboard);\n\n/***/ }),\n\n/***/ 828:\n/***/ (function(module) {\n\nvar DOCUMENT_NODE_TYPE = 9;\n\n/**\n * A polyfill for Element.matches()\n */\nif (typeof Element !== 'undefined' && !Element.prototype.matches) {\n var proto = Element.prototype;\n\n proto.matches = proto.matchesSelector ||\n proto.mozMatchesSelector ||\n proto.msMatchesSelector ||\n proto.oMatchesSelector ||\n proto.webkitMatchesSelector;\n}\n\n/**\n * Finds the closest parent that matches a selector.\n *\n * @param {Element} element\n * @param {String} selector\n * @return {Function}\n */\nfunction closest (element, selector) {\n while (element && element.nodeType !== DOCUMENT_NODE_TYPE) {\n if (typeof element.matches === 'function' &&\n element.matches(selector)) {\n return element;\n }\n element = element.parentNode;\n }\n}\n\nmodule.exports = closest;\n\n\n/***/ }),\n\n/***/ 438:\n/***/ (function(module, __unused_webpack_exports, __webpack_require__) {\n\nvar closest = __webpack_require__(828);\n\n/**\n * Delegates event to a selector.\n *\n * @param {Element} element\n * @param {String} selector\n * @param {String} type\n * @param {Function} callback\n * @param {Boolean} useCapture\n * @return {Object}\n */\nfunction _delegate(element, selector, type, callback, useCapture) {\n var listenerFn = listener.apply(this, arguments);\n\n element.addEventListener(type, listenerFn, useCapture);\n\n return {\n destroy: function() {\n element.removeEventListener(type, listenerFn, useCapture);\n }\n }\n}\n\n/**\n * Delegates event to a selector.\n *\n * @param {Element|String|Array} [elements]\n * @param {String} selector\n * @param {String} type\n * @param {Function} callback\n * @param {Boolean} useCapture\n * @return {Object}\n */\nfunction delegate(elements, selector, type, callback, useCapture) {\n // Handle the regular Element usage\n if (typeof elements.addEventListener === 'function') {\n return _delegate.apply(null, arguments);\n }\n\n // Handle Element-less usage, it defaults to global delegation\n if (typeof type === 'function') {\n // Use `document` as the first parameter, then apply arguments\n // This is a short way to .unshift `arguments` without running into deoptimizations\n return _delegate.bind(null, document).apply(null, arguments);\n }\n\n // Handle Selector-based usage\n if (typeof elements === 'string') {\n elements = document.querySelectorAll(elements);\n }\n\n // Handle Array-like based usage\n return Array.prototype.map.call(elements, function (element) {\n return _delegate(element, selector, type, callback, useCapture);\n });\n}\n\n/**\n * Finds closest match and invokes callback.\n *\n * @param {Element} element\n * @param {String} selector\n * @param {String} type\n * @param {Function} callback\n * @return {Function}\n */\nfunction listener(element, selector, type, callback) {\n return function(e) {\n e.delegateTarget = closest(e.target, selector);\n\n if (e.delegateTarget) {\n callback.call(element, e);\n }\n }\n}\n\nmodule.exports = delegate;\n\n\n/***/ }),\n\n/***/ 879:\n/***/ (function(__unused_webpack_module, exports) {\n\n/**\n * Check if argument is a HTML element.\n *\n * @param {Object} value\n * @return {Boolean}\n */\nexports.node = function(value) {\n return value !== undefined\n && value instanceof HTMLElement\n && value.nodeType === 1;\n};\n\n/**\n * Check if argument is a list of HTML elements.\n *\n * @param {Object} value\n * @return {Boolean}\n */\nexports.nodeList = function(value) {\n var type = Object.prototype.toString.call(value);\n\n return value !== undefined\n && (type === '[object NodeList]' || type === '[object HTMLCollection]')\n && ('length' in value)\n && (value.length === 0 || exports.node(value[0]));\n};\n\n/**\n * Check if argument is a string.\n *\n * @param {Object} value\n * @return {Boolean}\n */\nexports.string = function(value) {\n return typeof value === 'string'\n || value instanceof String;\n};\n\n/**\n * Check if argument is a function.\n *\n * @param {Object} value\n * @return {Boolean}\n */\nexports.fn = function(value) {\n var type = Object.prototype.toString.call(value);\n\n return type === '[object Function]';\n};\n\n\n/***/ }),\n\n/***/ 370:\n/***/ (function(module, __unused_webpack_exports, __webpack_require__) {\n\nvar is = __webpack_require__(879);\nvar delegate = __webpack_require__(438);\n\n/**\n * Validates all params and calls the right\n * listener function based on its target type.\n *\n * @param {String|HTMLElement|HTMLCollection|NodeList} target\n * @param {String} type\n * @param {Function} callback\n * @return {Object}\n */\nfunction listen(target, type, callback) {\n if (!target && !type && !callback) {\n throw new Error('Missing required arguments');\n }\n\n if (!is.string(type)) {\n throw new TypeError('Second argument must be a String');\n }\n\n if (!is.fn(callback)) {\n throw new TypeError('Third argument must be a Function');\n }\n\n if (is.node(target)) {\n return listenNode(target, type, callback);\n }\n else if (is.nodeList(target)) {\n return listenNodeList(target, type, callback);\n }\n else if (is.string(target)) {\n return listenSelector(target, type, callback);\n }\n else {\n throw new TypeError('First argument must be a String, HTMLElement, HTMLCollection, or NodeList');\n }\n}\n\n/**\n * Adds an event listener to a HTML element\n * and returns a remove listener function.\n *\n * @param {HTMLElement} node\n * @param {String} type\n * @param {Function} callback\n * @return {Object}\n */\nfunction listenNode(node, type, callback) {\n node.addEventListener(type, callback);\n\n return {\n destroy: function() {\n node.removeEventListener(type, callback);\n }\n }\n}\n\n/**\n * Add an event listener to a list of HTML elements\n * and returns a remove listener function.\n *\n * @param {NodeList|HTMLCollection} nodeList\n * @param {String} type\n * @param {Function} callback\n * @return {Object}\n */\nfunction listenNodeList(nodeList, type, callback) {\n Array.prototype.forEach.call(nodeList, function(node) {\n node.addEventListener(type, callback);\n });\n\n return {\n destroy: function() {\n Array.prototype.forEach.call(nodeList, function(node) {\n node.removeEventListener(type, callback);\n });\n }\n }\n}\n\n/**\n * Add an event listener to a selector\n * and returns a remove listener function.\n *\n * @param {String} selector\n * @param {String} type\n * @param {Function} callback\n * @return {Object}\n */\nfunction listenSelector(selector, type, callback) {\n return delegate(document.body, selector, type, callback);\n}\n\nmodule.exports = listen;\n\n\n/***/ }),\n\n/***/ 817:\n/***/ (function(module) {\n\nfunction select(element) {\n var selectedText;\n\n if (element.nodeName === 'SELECT') {\n element.focus();\n\n selectedText = element.value;\n }\n else if (element.nodeName === 'INPUT' || element.nodeName === 'TEXTAREA') {\n var isReadOnly = element.hasAttribute('readonly');\n\n if (!isReadOnly) {\n element.setAttribute('readonly', '');\n }\n\n element.select();\n element.setSelectionRange(0, element.value.length);\n\n if (!isReadOnly) {\n element.removeAttribute('readonly');\n }\n\n selectedText = element.value;\n }\n else {\n if (element.hasAttribute('contenteditable')) {\n element.focus();\n }\n\n var selection = window.getSelection();\n var range = document.createRange();\n\n range.selectNodeContents(element);\n selection.removeAllRanges();\n selection.addRange(range);\n\n selectedText = selection.toString();\n }\n\n return selectedText;\n}\n\nmodule.exports = select;\n\n\n/***/ }),\n\n/***/ 279:\n/***/ (function(module) {\n\nfunction E () {\n // Keep this empty so it's easier to inherit from\n // (via https://github.com/lipsmack from https://github.com/scottcorgan/tiny-emitter/issues/3)\n}\n\nE.prototype = {\n on: function (name, callback, ctx) {\n var e = this.e || (this.e = {});\n\n (e[name] || (e[name] = [])).push({\n fn: callback,\n ctx: ctx\n });\n\n return this;\n },\n\n once: function (name, callback, ctx) {\n var self = this;\n function listener () {\n self.off(name, listener);\n callback.apply(ctx, arguments);\n };\n\n listener._ = callback\n return this.on(name, listener, ctx);\n },\n\n emit: function (name) {\n var data = [].slice.call(arguments, 1);\n var evtArr = ((this.e || (this.e = {}))[name] || []).slice();\n var i = 0;\n var len = evtArr.length;\n\n for (i; i < len; i++) {\n evtArr[i].fn.apply(evtArr[i].ctx, data);\n }\n\n return this;\n },\n\n off: function (name, callback) {\n var e = this.e || (this.e = {});\n var evts = e[name];\n var liveEvents = [];\n\n if (evts && callback) {\n for (var i = 0, len = evts.length; i < len; i++) {\n if (evts[i].fn !== callback && evts[i].fn._ !== callback)\n liveEvents.push(evts[i]);\n }\n }\n\n // Remove event from queue to prevent memory leak\n // Suggested by https://github.com/lazd\n // Ref: https://github.com/scottcorgan/tiny-emitter/commit/c6ebfaa9bc973b33d110a84a307742b7cf94c953#commitcomment-5024910\n\n (liveEvents.length)\n ? e[name] = liveEvents\n : delete e[name];\n\n return this;\n }\n};\n\nmodule.exports = E;\nmodule.exports.TinyEmitter = E;\n\n\n/***/ })\n\n/******/ \t});\n/************************************************************************/\n/******/ \t// The module cache\n/******/ \tvar __webpack_module_cache__ = {};\n/******/ \t\n/******/ \t// The require function\n/******/ \tfunction __webpack_require__(moduleId) {\n/******/ \t\t// Check if module is in cache\n/******/ \t\tif(__webpack_module_cache__[moduleId]) {\n/******/ \t\t\treturn __webpack_module_cache__[moduleId].exports;\n/******/ \t\t}\n/******/ \t\t// Create a new module (and put it into the cache)\n/******/ \t\tvar module = __webpack_module_cache__[moduleId] = {\n/******/ \t\t\t// no module.id needed\n/******/ \t\t\t// no module.loaded needed\n/******/ \t\t\texports: {}\n/******/ \t\t};\n/******/ \t\n/******/ \t\t// Execute the module function\n/******/ \t\t__webpack_modules__[moduleId](module, module.exports, __webpack_require__);\n/******/ \t\n/******/ \t\t// Return the exports of the module\n/******/ \t\treturn module.exports;\n/******/ \t}\n/******/ \t\n/************************************************************************/\n/******/ \t/* webpack/runtime/compat get default export */\n/******/ \t!function() {\n/******/ \t\t// getDefaultExport function for compatibility with non-harmony modules\n/******/ \t\t__webpack_require__.n = function(module) {\n/******/ \t\t\tvar getter = module && module.__esModule ?\n/******/ \t\t\t\tfunction() { return module['default']; } :\n/******/ \t\t\t\tfunction() { return module; };\n/******/ \t\t\t__webpack_require__.d(getter, { a: getter });\n/******/ \t\t\treturn getter;\n/******/ \t\t};\n/******/ \t}();\n/******/ \t\n/******/ \t/* webpack/runtime/define property getters */\n/******/ \t!function() {\n/******/ \t\t// define getter functions for harmony exports\n/******/ \t\t__webpack_require__.d = function(exports, definition) {\n/******/ \t\t\tfor(var key in definition) {\n/******/ \t\t\t\tif(__webpack_require__.o(definition, key) && !__webpack_require__.o(exports, key)) {\n/******/ \t\t\t\t\tObject.defineProperty(exports, key, { enumerable: true, get: definition[key] });\n/******/ \t\t\t\t}\n/******/ \t\t\t}\n/******/ \t\t};\n/******/ \t}();\n/******/ \t\n/******/ \t/* webpack/runtime/hasOwnProperty shorthand */\n/******/ \t!function() {\n/******/ \t\t__webpack_require__.o = function(obj, prop) { return Object.prototype.hasOwnProperty.call(obj, prop); }\n/******/ \t}();\n/******/ \t\n/************************************************************************/\n/******/ \t// module exports must be returned from runtime so entry inlining is disabled\n/******/ \t// startup\n/******/ \t// Load entry module and return exports\n/******/ \treturn __webpack_require__(686);\n/******/ })()\n.default;\n});", "/*\n * Copyright (c) 2016-2024 Martin Donath \n *\n * Permission is hereby granted, free of charge, to any person obtaining a copy\n * of this software and associated documentation files (the \"Software\"), to\n * deal in the Software without restriction, including without limitation the\n * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or\n * sell copies of the Software, and to permit persons to whom the Software is\n * furnished to do so, subject to the following conditions:\n *\n * The above copyright notice and this permission notice shall be included in\n * all copies or substantial portions of the Software.\n *\n * THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL THE\n * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING\n * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS\n * IN THE SOFTWARE.\n */\n\nimport \"focus-visible\"\n\nimport {\n EMPTY,\n NEVER,\n Observable,\n Subject,\n defer,\n delay,\n filter,\n map,\n merge,\n mergeWith,\n shareReplay,\n switchMap\n} from \"rxjs\"\n\nimport { configuration, feature } from \"./_\"\nimport {\n at,\n getActiveElement,\n getOptionalElement,\n requestJSON,\n setLocation,\n setToggle,\n watchDocument,\n watchKeyboard,\n watchLocation,\n watchLocationTarget,\n watchMedia,\n watchPrint,\n watchScript,\n watchViewport\n} from \"./browser\"\nimport {\n getComponentElement,\n getComponentElements,\n mountAnnounce,\n mountBackToTop,\n mountConsent,\n mountContent,\n mountDialog,\n mountHeader,\n mountHeaderTitle,\n mountPalette,\n mountProgress,\n mountSearch,\n mountSearchHiglight,\n mountSidebar,\n mountSource,\n mountTableOfContents,\n mountTabs,\n watchHeader,\n watchMain\n} from \"./components\"\nimport {\n SearchIndex,\n setupClipboardJS,\n setupInstantNavigation,\n setupVersionSelector\n} from \"./integrations\"\nimport {\n patchEllipsis,\n patchIndeterminate,\n patchScrollfix,\n patchScrolllock\n} from \"./patches\"\nimport \"./polyfills\"\n\n/* ----------------------------------------------------------------------------\n * Functions - @todo refactor\n * ------------------------------------------------------------------------- */\n\n/**\n * Fetch search index\n *\n * @returns Search index observable\n */\nfunction fetchSearchIndex(): Observable {\n if (location.protocol === \"file:\") {\n return watchScript(\n `${new URL(\"search/search_index.js\", config.base)}`\n )\n .pipe(\n // @ts-ignore - @todo fix typings\n map(() => __index),\n shareReplay(1)\n )\n } else {\n return requestJSON(\n new URL(\"search/search_index.json\", config.base)\n )\n }\n}\n\n/* ----------------------------------------------------------------------------\n * Application\n * ------------------------------------------------------------------------- */\n\n/* Yay, JavaScript is available */\ndocument.documentElement.classList.remove(\"no-js\")\ndocument.documentElement.classList.add(\"js\")\n\n/* Set up navigation observables and subjects */\nconst document$ = watchDocument()\nconst location$ = watchLocation()\nconst target$ = watchLocationTarget(location$)\nconst keyboard$ = watchKeyboard()\n\n/* Set up media observables */\nconst viewport$ = watchViewport()\nconst tablet$ = watchMedia(\"(min-width: 960px)\")\nconst screen$ = watchMedia(\"(min-width: 1220px)\")\nconst print$ = watchPrint()\n\n/* Retrieve search index, if search is enabled */\nconst config = configuration()\nconst index$ = document.forms.namedItem(\"search\")\n ? fetchSearchIndex()\n : NEVER\n\n/* Set up Clipboard.js integration */\nconst alert$ = new Subject()\nsetupClipboardJS({ alert$ })\n\n/* Set up progress indicator */\nconst progress$ = new Subject()\n\n/* Set up instant navigation, if enabled */\nif (feature(\"navigation.instant\"))\n setupInstantNavigation({ location$, viewport$, progress$ })\n .subscribe(document$)\n\n/* Set up version selector */\nif (config.version?.provider === \"mike\")\n setupVersionSelector({ document$ })\n\n/* Always close drawer and search on navigation */\nmerge(location$, target$)\n .pipe(\n delay(125)\n )\n .subscribe(() => {\n setToggle(\"drawer\", false)\n setToggle(\"search\", false)\n })\n\n/* Set up global keyboard handlers */\nkeyboard$\n .pipe(\n filter(({ mode }) => mode === \"global\")\n )\n .subscribe(key => {\n switch (key.type) {\n\n /* Go to previous page */\n case \"p\":\n case \",\":\n const prev = getOptionalElement(\"link[rel=prev]\")\n if (typeof prev !== \"undefined\")\n setLocation(prev)\n break\n\n /* Go to next page */\n case \"n\":\n case \".\":\n const next = getOptionalElement(\"link[rel=next]\")\n if (typeof next !== \"undefined\")\n setLocation(next)\n break\n\n /* Expand navigation, see https://bit.ly/3ZjG5io */\n case \"Enter\":\n const active = getActiveElement()\n if (active instanceof HTMLLabelElement)\n active.click()\n }\n })\n\n/* Set up patches */\npatchEllipsis({ viewport$, document$ })\npatchIndeterminate({ document$, tablet$ })\npatchScrollfix({ document$ })\npatchScrolllock({ viewport$, tablet$ })\n\n/* Set up header and main area observable */\nconst header$ = watchHeader(getComponentElement(\"header\"), { viewport$ })\nconst main$ = document$\n .pipe(\n map(() => getComponentElement(\"main\")),\n switchMap(el => watchMain(el, { viewport$, header$ })),\n shareReplay(1)\n )\n\n/* Set up control component observables */\nconst control$ = merge(\n\n /* Consent */\n ...getComponentElements(\"consent\")\n .map(el => mountConsent(el, { target$ })),\n\n /* Dialog */\n ...getComponentElements(\"dialog\")\n .map(el => mountDialog(el, { alert$ })),\n\n /* Color palette */\n ...getComponentElements(\"palette\")\n .map(el => mountPalette(el)),\n\n /* Progress bar */\n ...getComponentElements(\"progress\")\n .map(el => mountProgress(el, { progress$ })),\n\n /* Search */\n ...getComponentElements(\"search\")\n .map(el => mountSearch(el, { index$, keyboard$ })),\n\n /* Repository information */\n ...getComponentElements(\"source\")\n .map(el => mountSource(el))\n)\n\n/* Set up content component observables */\nconst content$ = defer(() => merge(\n\n /* Announcement bar */\n ...getComponentElements(\"announce\")\n .map(el => mountAnnounce(el)),\n\n /* Content */\n ...getComponentElements(\"content\")\n .map(el => mountContent(el, { viewport$, target$, print$ })),\n\n /* Search highlighting */\n ...getComponentElements(\"content\")\n .map(el => feature(\"search.highlight\")\n ? mountSearchHiglight(el, { index$, location$ })\n : EMPTY\n ),\n\n /* Header */\n ...getComponentElements(\"header\")\n .map(el => mountHeader(el, { viewport$, header$, main$ })),\n\n /* Header title */\n ...getComponentElements(\"header-title\")\n .map(el => mountHeaderTitle(el, { viewport$, header$ })),\n\n /* Sidebar */\n ...getComponentElements(\"sidebar\")\n .map(el => el.getAttribute(\"data-md-type\") === \"navigation\"\n ? at(screen$, () => mountSidebar(el, { viewport$, header$, main$ }))\n : at(tablet$, () => mountSidebar(el, { viewport$, header$, main$ }))\n ),\n\n /* Navigation tabs */\n ...getComponentElements(\"tabs\")\n .map(el => mountTabs(el, { viewport$, header$ })),\n\n /* Table of contents */\n ...getComponentElements(\"toc\")\n .map(el => mountTableOfContents(el, {\n viewport$, header$, main$, target$\n })),\n\n /* Back-to-top button */\n ...getComponentElements(\"top\")\n .map(el => mountBackToTop(el, { viewport$, header$, main$, target$ }))\n))\n\n/* Set up component observables */\nconst component$ = document$\n .pipe(\n switchMap(() => content$),\n mergeWith(control$),\n shareReplay(1)\n )\n\n/* Subscribe to all components */\ncomponent$.subscribe()\n\n/* ----------------------------------------------------------------------------\n * Exports\n * ------------------------------------------------------------------------- */\n\nwindow.document$ = document$ /* Document observable */\nwindow.location$ = location$ /* Location subject */\nwindow.target$ = target$ /* Location target observable */\nwindow.keyboard$ = keyboard$ /* Keyboard observable */\nwindow.viewport$ = viewport$ /* Viewport observable */\nwindow.tablet$ = tablet$ /* Media tablet observable */\nwindow.screen$ = screen$ /* Media screen observable */\nwindow.print$ = print$ /* Media print observable */\nwindow.alert$ = alert$ /* Alert subject */\nwindow.progress$ = progress$ /* Progress indicator subject */\nwindow.component$ = component$ /* Component observable */\n", "/******************************************************************************\nCopyright (c) Microsoft Corporation.\n\nPermission to use, copy, modify, and/or distribute this software for any\npurpose with or without fee is hereby granted.\n\nTHE SOFTWARE IS PROVIDED \"AS IS\" AND THE AUTHOR DISCLAIMS ALL WARRANTIES WITH\nREGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY\nAND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT,\nINDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM\nLOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR\nOTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR\nPERFORMANCE OF THIS SOFTWARE.\n***************************************************************************** */\n/* global Reflect, Promise, SuppressedError, Symbol, Iterator */\n\nvar extendStatics = function(d, b) {\n extendStatics = Object.setPrototypeOf ||\n ({ __proto__: [] } instanceof Array && function (d, b) { d.__proto__ = b; }) ||\n function (d, b) { for (var p in b) if (Object.prototype.hasOwnProperty.call(b, p)) d[p] = b[p]; };\n return extendStatics(d, b);\n};\n\nexport function __extends(d, b) {\n if (typeof b !== \"function\" && b !== null)\n throw new TypeError(\"Class extends value \" + String(b) + \" is not a constructor or null\");\n extendStatics(d, b);\n function __() { this.constructor = d; }\n d.prototype = b === null ? Object.create(b) : (__.prototype = b.prototype, new __());\n}\n\nexport var __assign = function() {\n __assign = Object.assign || function __assign(t) {\n for (var s, i = 1, n = arguments.length; i < n; i++) {\n s = arguments[i];\n for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p)) t[p] = s[p];\n }\n return t;\n }\n return __assign.apply(this, arguments);\n}\n\nexport function __rest(s, e) {\n var t = {};\n for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p) && e.indexOf(p) < 0)\n t[p] = s[p];\n if (s != null && typeof Object.getOwnPropertySymbols === \"function\")\n for (var i = 0, p = Object.getOwnPropertySymbols(s); i < p.length; i++) {\n if (e.indexOf(p[i]) < 0 && Object.prototype.propertyIsEnumerable.call(s, p[i]))\n t[p[i]] = s[p[i]];\n }\n return t;\n}\n\nexport function __decorate(decorators, target, key, desc) {\n var c = arguments.length, r = c < 3 ? target : desc === null ? desc = Object.getOwnPropertyDescriptor(target, key) : desc, d;\n if (typeof Reflect === \"object\" && typeof Reflect.decorate === \"function\") r = Reflect.decorate(decorators, target, key, desc);\n else for (var i = decorators.length - 1; i >= 0; i--) if (d = decorators[i]) r = (c < 3 ? d(r) : c > 3 ? d(target, key, r) : d(target, key)) || r;\n return c > 3 && r && Object.defineProperty(target, key, r), r;\n}\n\nexport function __param(paramIndex, decorator) {\n return function (target, key) { decorator(target, key, paramIndex); }\n}\n\nexport function __esDecorate(ctor, descriptorIn, decorators, contextIn, initializers, extraInitializers) {\n function accept(f) { if (f !== void 0 && typeof f !== \"function\") throw new TypeError(\"Function expected\"); return f; }\n var kind = contextIn.kind, key = kind === \"getter\" ? \"get\" : kind === \"setter\" ? \"set\" : \"value\";\n var target = !descriptorIn && ctor ? contextIn[\"static\"] ? ctor : ctor.prototype : null;\n var descriptor = descriptorIn || (target ? Object.getOwnPropertyDescriptor(target, contextIn.name) : {});\n var _, done = false;\n for (var i = decorators.length - 1; i >= 0; i--) {\n var context = {};\n for (var p in contextIn) context[p] = p === \"access\" ? {} : contextIn[p];\n for (var p in contextIn.access) context.access[p] = contextIn.access[p];\n context.addInitializer = function (f) { if (done) throw new TypeError(\"Cannot add initializers after decoration has completed\"); extraInitializers.push(accept(f || null)); };\n var result = (0, decorators[i])(kind === \"accessor\" ? { get: descriptor.get, set: descriptor.set } : descriptor[key], context);\n if (kind === \"accessor\") {\n if (result === void 0) continue;\n if (result === null || typeof result !== \"object\") throw new TypeError(\"Object expected\");\n if (_ = accept(result.get)) descriptor.get = _;\n if (_ = accept(result.set)) descriptor.set = _;\n if (_ = accept(result.init)) initializers.unshift(_);\n }\n else if (_ = accept(result)) {\n if (kind === \"field\") initializers.unshift(_);\n else descriptor[key] = _;\n }\n }\n if (target) Object.defineProperty(target, contextIn.name, descriptor);\n done = true;\n};\n\nexport function __runInitializers(thisArg, initializers, value) {\n var useValue = arguments.length > 2;\n for (var i = 0; i < initializers.length; i++) {\n value = useValue ? initializers[i].call(thisArg, value) : initializers[i].call(thisArg);\n }\n return useValue ? value : void 0;\n};\n\nexport function __propKey(x) {\n return typeof x === \"symbol\" ? x : \"\".concat(x);\n};\n\nexport function __setFunctionName(f, name, prefix) {\n if (typeof name === \"symbol\") name = name.description ? \"[\".concat(name.description, \"]\") : \"\";\n return Object.defineProperty(f, \"name\", { configurable: true, value: prefix ? \"\".concat(prefix, \" \", name) : name });\n};\n\nexport function __metadata(metadataKey, metadataValue) {\n if (typeof Reflect === \"object\" && typeof Reflect.metadata === \"function\") return Reflect.metadata(metadataKey, metadataValue);\n}\n\nexport function __awaiter(thisArg, _arguments, P, generator) {\n function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }\n return new (P || (P = Promise))(function (resolve, reject) {\n function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }\n function rejected(value) { try { step(generator[\"throw\"](value)); } catch (e) { reject(e); } }\n function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }\n step((generator = generator.apply(thisArg, _arguments || [])).next());\n });\n}\n\nexport function __generator(thisArg, body) {\n var _ = { label: 0, sent: function() { if (t[0] & 1) throw t[1]; return t[1]; }, trys: [], ops: [] }, f, y, t, g = Object.create((typeof Iterator === \"function\" ? Iterator : Object).prototype);\n return g.next = verb(0), g[\"throw\"] = verb(1), g[\"return\"] = verb(2), typeof Symbol === \"function\" && (g[Symbol.iterator] = function() { return this; }), g;\n function verb(n) { return function (v) { return step([n, v]); }; }\n function step(op) {\n if (f) throw new TypeError(\"Generator is already executing.\");\n while (g && (g = 0, op[0] && (_ = 0)), _) try {\n if (f = 1, y && (t = op[0] & 2 ? y[\"return\"] : op[0] ? y[\"throw\"] || ((t = y[\"return\"]) && t.call(y), 0) : y.next) && !(t = t.call(y, op[1])).done) return t;\n if (y = 0, t) op = [op[0] & 2, t.value];\n switch (op[0]) {\n case 0: case 1: t = op; break;\n case 4: _.label++; return { value: op[1], done: false };\n case 5: _.label++; y = op[1]; op = [0]; continue;\n case 7: op = _.ops.pop(); _.trys.pop(); continue;\n default:\n if (!(t = _.trys, t = t.length > 0 && t[t.length - 1]) && (op[0] === 6 || op[0] === 2)) { _ = 0; continue; }\n if (op[0] === 3 && (!t || (op[1] > t[0] && op[1] < t[3]))) { _.label = op[1]; break; }\n if (op[0] === 6 && _.label < t[1]) { _.label = t[1]; t = op; break; }\n if (t && _.label < t[2]) { _.label = t[2]; _.ops.push(op); break; }\n if (t[2]) _.ops.pop();\n _.trys.pop(); continue;\n }\n op = body.call(thisArg, _);\n } catch (e) { op = [6, e]; y = 0; } finally { f = t = 0; }\n if (op[0] & 5) throw op[1]; return { value: op[0] ? op[1] : void 0, done: true };\n }\n}\n\nexport var __createBinding = Object.create ? (function(o, m, k, k2) {\n if (k2 === undefined) k2 = k;\n var desc = Object.getOwnPropertyDescriptor(m, k);\n if (!desc || (\"get\" in desc ? !m.__esModule : desc.writable || desc.configurable)) {\n desc = { enumerable: true, get: function() { return m[k]; } };\n }\n Object.defineProperty(o, k2, desc);\n}) : (function(o, m, k, k2) {\n if (k2 === undefined) k2 = k;\n o[k2] = m[k];\n});\n\nexport function __exportStar(m, o) {\n for (var p in m) if (p !== \"default\" && !Object.prototype.hasOwnProperty.call(o, p)) __createBinding(o, m, p);\n}\n\nexport function __values(o) {\n var s = typeof Symbol === \"function\" && Symbol.iterator, m = s && o[s], i = 0;\n if (m) return m.call(o);\n if (o && typeof o.length === \"number\") return {\n next: function () {\n if (o && i >= o.length) o = void 0;\n return { value: o && o[i++], done: !o };\n }\n };\n throw new TypeError(s ? \"Object is not iterable.\" : \"Symbol.iterator is not defined.\");\n}\n\nexport function __read(o, n) {\n var m = typeof Symbol === \"function\" && o[Symbol.iterator];\n if (!m) return o;\n var i = m.call(o), r, ar = [], e;\n try {\n while ((n === void 0 || n-- > 0) && !(r = i.next()).done) ar.push(r.value);\n }\n catch (error) { e = { error: error }; }\n finally {\n try {\n if (r && !r.done && (m = i[\"return\"])) m.call(i);\n }\n finally { if (e) throw e.error; }\n }\n return ar;\n}\n\n/** @deprecated */\nexport function __spread() {\n for (var ar = [], i = 0; i < arguments.length; i++)\n ar = ar.concat(__read(arguments[i]));\n return ar;\n}\n\n/** @deprecated */\nexport function __spreadArrays() {\n for (var s = 0, i = 0, il = arguments.length; i < il; i++) s += arguments[i].length;\n for (var r = Array(s), k = 0, i = 0; i < il; i++)\n for (var a = arguments[i], j = 0, jl = a.length; j < jl; j++, k++)\n r[k] = a[j];\n return r;\n}\n\nexport function __spreadArray(to, from, pack) {\n if (pack || arguments.length === 2) for (var i = 0, l = from.length, ar; i < l; i++) {\n if (ar || !(i in from)) {\n if (!ar) ar = Array.prototype.slice.call(from, 0, i);\n ar[i] = from[i];\n }\n }\n return to.concat(ar || Array.prototype.slice.call(from));\n}\n\nexport function __await(v) {\n return this instanceof __await ? (this.v = v, this) : new __await(v);\n}\n\nexport function __asyncGenerator(thisArg, _arguments, generator) {\n if (!Symbol.asyncIterator) throw new TypeError(\"Symbol.asyncIterator is not defined.\");\n var g = generator.apply(thisArg, _arguments || []), i, q = [];\n return i = Object.create((typeof AsyncIterator === \"function\" ? AsyncIterator : Object).prototype), verb(\"next\"), verb(\"throw\"), verb(\"return\", awaitReturn), i[Symbol.asyncIterator] = function () { return this; }, i;\n function awaitReturn(f) { return function (v) { return Promise.resolve(v).then(f, reject); }; }\n function verb(n, f) { if (g[n]) { i[n] = function (v) { return new Promise(function (a, b) { q.push([n, v, a, b]) > 1 || resume(n, v); }); }; if (f) i[n] = f(i[n]); } }\n function resume(n, v) { try { step(g[n](v)); } catch (e) { settle(q[0][3], e); } }\n function step(r) { r.value instanceof __await ? Promise.resolve(r.value.v).then(fulfill, reject) : settle(q[0][2], r); }\n function fulfill(value) { resume(\"next\", value); }\n function reject(value) { resume(\"throw\", value); }\n function settle(f, v) { if (f(v), q.shift(), q.length) resume(q[0][0], q[0][1]); }\n}\n\nexport function __asyncDelegator(o) {\n var i, p;\n return i = {}, verb(\"next\"), verb(\"throw\", function (e) { throw e; }), verb(\"return\"), i[Symbol.iterator] = function () { return this; }, i;\n function verb(n, f) { i[n] = o[n] ? function (v) { return (p = !p) ? { value: __await(o[n](v)), done: false } : f ? f(v) : v; } : f; }\n}\n\nexport function __asyncValues(o) {\n if (!Symbol.asyncIterator) throw new TypeError(\"Symbol.asyncIterator is not defined.\");\n var m = o[Symbol.asyncIterator], i;\n return m ? m.call(o) : (o = typeof __values === \"function\" ? __values(o) : o[Symbol.iterator](), i = {}, verb(\"next\"), verb(\"throw\"), verb(\"return\"), i[Symbol.asyncIterator] = function () { return this; }, i);\n function verb(n) { i[n] = o[n] && function (v) { return new Promise(function (resolve, reject) { v = o[n](v), settle(resolve, reject, v.done, v.value); }); }; }\n function settle(resolve, reject, d, v) { Promise.resolve(v).then(function(v) { resolve({ value: v, done: d }); }, reject); }\n}\n\nexport function __makeTemplateObject(cooked, raw) {\n if (Object.defineProperty) { Object.defineProperty(cooked, \"raw\", { value: raw }); } else { cooked.raw = raw; }\n return cooked;\n};\n\nvar __setModuleDefault = Object.create ? (function(o, v) {\n Object.defineProperty(o, \"default\", { enumerable: true, value: v });\n}) : function(o, v) {\n o[\"default\"] = v;\n};\n\nexport function __importStar(mod) {\n if (mod && mod.__esModule) return mod;\n var result = {};\n if (mod != null) for (var k in mod) if (k !== \"default\" && Object.prototype.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);\n __setModuleDefault(result, mod);\n return result;\n}\n\nexport function __importDefault(mod) {\n return (mod && mod.__esModule) ? mod : { default: mod };\n}\n\nexport function __classPrivateFieldGet(receiver, state, kind, f) {\n if (kind === \"a\" && !f) throw new TypeError(\"Private accessor was defined without a getter\");\n if (typeof state === \"function\" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError(\"Cannot read private member from an object whose class did not declare it\");\n return kind === \"m\" ? f : kind === \"a\" ? f.call(receiver) : f ? f.value : state.get(receiver);\n}\n\nexport function __classPrivateFieldSet(receiver, state, value, kind, f) {\n if (kind === \"m\") throw new TypeError(\"Private method is not writable\");\n if (kind === \"a\" && !f) throw new TypeError(\"Private accessor was defined without a setter\");\n if (typeof state === \"function\" ? receiver !== state || !f : !state.has(receiver)) throw new TypeError(\"Cannot write private member to an object whose class did not declare it\");\n return (kind === \"a\" ? f.call(receiver, value) : f ? f.value = value : state.set(receiver, value)), value;\n}\n\nexport function __classPrivateFieldIn(state, receiver) {\n if (receiver === null || (typeof receiver !== \"object\" && typeof receiver !== \"function\")) throw new TypeError(\"Cannot use 'in' operator on non-object\");\n return typeof state === \"function\" ? receiver === state : state.has(receiver);\n}\n\nexport function __addDisposableResource(env, value, async) {\n if (value !== null && value !== void 0) {\n if (typeof value !== \"object\" && typeof value !== \"function\") throw new TypeError(\"Object expected.\");\n var dispose, inner;\n if (async) {\n if (!Symbol.asyncDispose) throw new TypeError(\"Symbol.asyncDispose is not defined.\");\n dispose = value[Symbol.asyncDispose];\n }\n if (dispose === void 0) {\n if (!Symbol.dispose) throw new TypeError(\"Symbol.dispose is not defined.\");\n dispose = value[Symbol.dispose];\n if (async) inner = dispose;\n }\n if (typeof dispose !== \"function\") throw new TypeError(\"Object not disposable.\");\n if (inner) dispose = function() { try { inner.call(this); } catch (e) { return Promise.reject(e); } };\n env.stack.push({ value: value, dispose: dispose, async: async });\n }\n else if (async) {\n env.stack.push({ async: true });\n }\n return value;\n}\n\nvar _SuppressedError = typeof SuppressedError === \"function\" ? SuppressedError : function (error, suppressed, message) {\n var e = new Error(message);\n return e.name = \"SuppressedError\", e.error = error, e.suppressed = suppressed, e;\n};\n\nexport function __disposeResources(env) {\n function fail(e) {\n env.error = env.hasError ? new _SuppressedError(e, env.error, \"An error was suppressed during disposal.\") : e;\n env.hasError = true;\n }\n var r, s = 0;\n function next() {\n while (r = env.stack.pop()) {\n try {\n if (!r.async && s === 1) return s = 0, env.stack.push(r), Promise.resolve().then(next);\n if (r.dispose) {\n var result = r.dispose.call(r.value);\n if (r.async) return s |= 2, Promise.resolve(result).then(next, function(e) { fail(e); return next(); });\n }\n else s |= 1;\n }\n catch (e) {\n fail(e);\n }\n }\n if (s === 1) return env.hasError ? Promise.reject(env.error) : Promise.resolve();\n if (env.hasError) throw env.error;\n }\n return next();\n}\n\nexport default {\n __extends,\n __assign,\n __rest,\n __decorate,\n __param,\n __metadata,\n __awaiter,\n __generator,\n __createBinding,\n __exportStar,\n __values,\n __read,\n __spread,\n __spreadArrays,\n __spreadArray,\n __await,\n __asyncGenerator,\n __asyncDelegator,\n __asyncValues,\n __makeTemplateObject,\n __importStar,\n __importDefault,\n __classPrivateFieldGet,\n __classPrivateFieldSet,\n __classPrivateFieldIn,\n __addDisposableResource,\n __disposeResources,\n};\n", "/**\n * Returns true if the object is a function.\n * @param value The value to check\n */\nexport function isFunction(value: any): value is (...args: any[]) => any {\n return typeof value === 'function';\n}\n", "/**\n * Used to create Error subclasses until the community moves away from ES5.\n *\n * This is because compiling from TypeScript down to ES5 has issues with subclassing Errors\n * as well as other built-in types: https://github.com/Microsoft/TypeScript/issues/12123\n *\n * @param createImpl A factory function to create the actual constructor implementation. The returned\n * function should be a named function that calls `_super` internally.\n */\nexport function createErrorClass(createImpl: (_super: any) => any): T {\n const _super = (instance: any) => {\n Error.call(instance);\n instance.stack = new Error().stack;\n };\n\n const ctorFunc = createImpl(_super);\n ctorFunc.prototype = Object.create(Error.prototype);\n ctorFunc.prototype.constructor = ctorFunc;\n return ctorFunc;\n}\n", "import { createErrorClass } from './createErrorClass';\n\nexport interface UnsubscriptionError extends Error {\n readonly errors: any[];\n}\n\nexport interface UnsubscriptionErrorCtor {\n /**\n * @deprecated Internal implementation detail. Do not construct error instances.\n * Cannot be tagged as internal: https://github.com/ReactiveX/rxjs/issues/6269\n */\n new (errors: any[]): UnsubscriptionError;\n}\n\n/**\n * An error thrown when one or more errors have occurred during the\n * `unsubscribe` of a {@link Subscription}.\n */\nexport const UnsubscriptionError: UnsubscriptionErrorCtor = createErrorClass(\n (_super) =>\n function UnsubscriptionErrorImpl(this: any, errors: (Error | string)[]) {\n _super(this);\n this.message = errors\n ? `${errors.length} errors occurred during unsubscription:\n${errors.map((err, i) => `${i + 1}) ${err.toString()}`).join('\\n ')}`\n : '';\n this.name = 'UnsubscriptionError';\n this.errors = errors;\n }\n);\n", "/**\n * Removes an item from an array, mutating it.\n * @param arr The array to remove the item from\n * @param item The item to remove\n */\nexport function arrRemove(arr: T[] | undefined | null, item: T) {\n if (arr) {\n const index = arr.indexOf(item);\n 0 <= index && arr.splice(index, 1);\n }\n}\n", "import { isFunction } from './util/isFunction';\nimport { UnsubscriptionError } from './util/UnsubscriptionError';\nimport { SubscriptionLike, TeardownLogic, Unsubscribable } from './types';\nimport { arrRemove } from './util/arrRemove';\n\n/**\n * Represents a disposable resource, such as the execution of an Observable. A\n * Subscription has one important method, `unsubscribe`, that takes no argument\n * and just disposes the resource held by the subscription.\n *\n * Additionally, subscriptions may be grouped together through the `add()`\n * method, which will attach a child Subscription to the current Subscription.\n * When a Subscription is unsubscribed, all its children (and its grandchildren)\n * will be unsubscribed as well.\n *\n * @class Subscription\n */\nexport class Subscription implements SubscriptionLike {\n /** @nocollapse */\n public static EMPTY = (() => {\n const empty = new Subscription();\n empty.closed = true;\n return empty;\n })();\n\n /**\n * A flag to indicate whether this Subscription has already been unsubscribed.\n */\n public closed = false;\n\n private _parentage: Subscription[] | Subscription | null = null;\n\n /**\n * The list of registered finalizers to execute upon unsubscription. Adding and removing from this\n * list occurs in the {@link #add} and {@link #remove} methods.\n */\n private _finalizers: Exclude[] | null = null;\n\n /**\n * @param initialTeardown A function executed first as part of the finalization\n * process that is kicked off when {@link #unsubscribe} is called.\n */\n constructor(private initialTeardown?: () => void) {}\n\n /**\n * Disposes the resources held by the subscription. May, for instance, cancel\n * an ongoing Observable execution or cancel any other type of work that\n * started when the Subscription was created.\n * @return {void}\n */\n unsubscribe(): void {\n let errors: any[] | undefined;\n\n if (!this.closed) {\n this.closed = true;\n\n // Remove this from it's parents.\n const { _parentage } = this;\n if (_parentage) {\n this._parentage = null;\n if (Array.isArray(_parentage)) {\n for (const parent of _parentage) {\n parent.remove(this);\n }\n } else {\n _parentage.remove(this);\n }\n }\n\n const { initialTeardown: initialFinalizer } = this;\n if (isFunction(initialFinalizer)) {\n try {\n initialFinalizer();\n } catch (e) {\n errors = e instanceof UnsubscriptionError ? e.errors : [e];\n }\n }\n\n const { _finalizers } = this;\n if (_finalizers) {\n this._finalizers = null;\n for (const finalizer of _finalizers) {\n try {\n execFinalizer(finalizer);\n } catch (err) {\n errors = errors ?? [];\n if (err instanceof UnsubscriptionError) {\n errors = [...errors, ...err.errors];\n } else {\n errors.push(err);\n }\n }\n }\n }\n\n if (errors) {\n throw new UnsubscriptionError(errors);\n }\n }\n }\n\n /**\n * Adds a finalizer to this subscription, so that finalization will be unsubscribed/called\n * when this subscription is unsubscribed. If this subscription is already {@link #closed},\n * because it has already been unsubscribed, then whatever finalizer is passed to it\n * will automatically be executed (unless the finalizer itself is also a closed subscription).\n *\n * Closed Subscriptions cannot be added as finalizers to any subscription. Adding a closed\n * subscription to a any subscription will result in no operation. (A noop).\n *\n * Adding a subscription to itself, or adding `null` or `undefined` will not perform any\n * operation at all. (A noop).\n *\n * `Subscription` instances that are added to this instance will automatically remove themselves\n * if they are unsubscribed. Functions and {@link Unsubscribable} objects that you wish to remove\n * will need to be removed manually with {@link #remove}\n *\n * @param teardown The finalization logic to add to this subscription.\n */\n add(teardown: TeardownLogic): void {\n // Only add the finalizer if it's not undefined\n // and don't add a subscription to itself.\n if (teardown && teardown !== this) {\n if (this.closed) {\n // If this subscription is already closed,\n // execute whatever finalizer is handed to it automatically.\n execFinalizer(teardown);\n } else {\n if (teardown instanceof Subscription) {\n // We don't add closed subscriptions, and we don't add the same subscription\n // twice. Subscription unsubscribe is idempotent.\n if (teardown.closed || teardown._hasParent(this)) {\n return;\n }\n teardown._addParent(this);\n }\n (this._finalizers = this._finalizers ?? []).push(teardown);\n }\n }\n }\n\n /**\n * Checks to see if a this subscription already has a particular parent.\n * This will signal that this subscription has already been added to the parent in question.\n * @param parent the parent to check for\n */\n private _hasParent(parent: Subscription) {\n const { _parentage } = this;\n return _parentage === parent || (Array.isArray(_parentage) && _parentage.includes(parent));\n }\n\n /**\n * Adds a parent to this subscription so it can be removed from the parent if it\n * unsubscribes on it's own.\n *\n * NOTE: THIS ASSUMES THAT {@link _hasParent} HAS ALREADY BEEN CHECKED.\n * @param parent The parent subscription to add\n */\n private _addParent(parent: Subscription) {\n const { _parentage } = this;\n this._parentage = Array.isArray(_parentage) ? (_parentage.push(parent), _parentage) : _parentage ? [_parentage, parent] : parent;\n }\n\n /**\n * Called on a child when it is removed via {@link #remove}.\n * @param parent The parent to remove\n */\n private _removeParent(parent: Subscription) {\n const { _parentage } = this;\n if (_parentage === parent) {\n this._parentage = null;\n } else if (Array.isArray(_parentage)) {\n arrRemove(_parentage, parent);\n }\n }\n\n /**\n * Removes a finalizer from this subscription that was previously added with the {@link #add} method.\n *\n * Note that `Subscription` instances, when unsubscribed, will automatically remove themselves\n * from every other `Subscription` they have been added to. This means that using the `remove` method\n * is not a common thing and should be used thoughtfully.\n *\n * If you add the same finalizer instance of a function or an unsubscribable object to a `Subscription` instance\n * more than once, you will need to call `remove` the same number of times to remove all instances.\n *\n * All finalizer instances are removed to free up memory upon unsubscription.\n *\n * @param teardown The finalizer to remove from this subscription\n */\n remove(teardown: Exclude): void {\n const { _finalizers } = this;\n _finalizers && arrRemove(_finalizers, teardown);\n\n if (teardown instanceof Subscription) {\n teardown._removeParent(this);\n }\n }\n}\n\nexport const EMPTY_SUBSCRIPTION = Subscription.EMPTY;\n\nexport function isSubscription(value: any): value is Subscription {\n return (\n value instanceof Subscription ||\n (value && 'closed' in value && isFunction(value.remove) && isFunction(value.add) && isFunction(value.unsubscribe))\n );\n}\n\nfunction execFinalizer(finalizer: Unsubscribable | (() => void)) {\n if (isFunction(finalizer)) {\n finalizer();\n } else {\n finalizer.unsubscribe();\n }\n}\n", "import { Subscriber } from './Subscriber';\nimport { ObservableNotification } from './types';\n\n/**\n * The {@link GlobalConfig} object for RxJS. It is used to configure things\n * like how to react on unhandled errors.\n */\nexport const config: GlobalConfig = {\n onUnhandledError: null,\n onStoppedNotification: null,\n Promise: undefined,\n useDeprecatedSynchronousErrorHandling: false,\n useDeprecatedNextContext: false,\n};\n\n/**\n * The global configuration object for RxJS, used to configure things\n * like how to react on unhandled errors. Accessible via {@link config}\n * object.\n */\nexport interface GlobalConfig {\n /**\n * A registration point for unhandled errors from RxJS. These are errors that\n * cannot were not handled by consuming code in the usual subscription path. For\n * example, if you have this configured, and you subscribe to an observable without\n * providing an error handler, errors from that subscription will end up here. This\n * will _always_ be called asynchronously on another job in the runtime. This is because\n * we do not want errors thrown in this user-configured handler to interfere with the\n * behavior of the library.\n */\n onUnhandledError: ((err: any) => void) | null;\n\n /**\n * A registration point for notifications that cannot be sent to subscribers because they\n * have completed, errored or have been explicitly unsubscribed. By default, next, complete\n * and error notifications sent to stopped subscribers are noops. However, sometimes callers\n * might want a different behavior. For example, with sources that attempt to report errors\n * to stopped subscribers, a caller can configure RxJS to throw an unhandled error instead.\n * This will _always_ be called asynchronously on another job in the runtime. This is because\n * we do not want errors thrown in this user-configured handler to interfere with the\n * behavior of the library.\n */\n onStoppedNotification: ((notification: ObservableNotification, subscriber: Subscriber) => void) | null;\n\n /**\n * The promise constructor used by default for {@link Observable#toPromise toPromise} and {@link Observable#forEach forEach}\n * methods.\n *\n * @deprecated As of version 8, RxJS will no longer support this sort of injection of a\n * Promise constructor. If you need a Promise implementation other than native promises,\n * please polyfill/patch Promise as you see appropriate. Will be removed in v8.\n */\n Promise?: PromiseConstructorLike;\n\n /**\n * If true, turns on synchronous error rethrowing, which is a deprecated behavior\n * in v6 and higher. This behavior enables bad patterns like wrapping a subscribe\n * call in a try/catch block. It also enables producer interference, a nasty bug\n * where a multicast can be broken for all observers by a downstream consumer with\n * an unhandled error. DO NOT USE THIS FLAG UNLESS IT'S NEEDED TO BUY TIME\n * FOR MIGRATION REASONS.\n *\n * @deprecated As of version 8, RxJS will no longer support synchronous throwing\n * of unhandled errors. All errors will be thrown on a separate call stack to prevent bad\n * behaviors described above. Will be removed in v8.\n */\n useDeprecatedSynchronousErrorHandling: boolean;\n\n /**\n * If true, enables an as-of-yet undocumented feature from v5: The ability to access\n * `unsubscribe()` via `this` context in `next` functions created in observers passed\n * to `subscribe`.\n *\n * This is being removed because the performance was severely problematic, and it could also cause\n * issues when types other than POJOs are passed to subscribe as subscribers, as they will likely have\n * their `this` context overwritten.\n *\n * @deprecated As of version 8, RxJS will no longer support altering the\n * context of next functions provided as part of an observer to Subscribe. Instead,\n * you will have access to a subscription or a signal or token that will allow you to do things like\n * unsubscribe and test closed status. Will be removed in v8.\n */\n useDeprecatedNextContext: boolean;\n}\n", "import type { TimerHandle } from './timerHandle';\ntype SetTimeoutFunction = (handler: () => void, timeout?: number, ...args: any[]) => TimerHandle;\ntype ClearTimeoutFunction = (handle: TimerHandle) => void;\n\ninterface TimeoutProvider {\n setTimeout: SetTimeoutFunction;\n clearTimeout: ClearTimeoutFunction;\n delegate:\n | {\n setTimeout: SetTimeoutFunction;\n clearTimeout: ClearTimeoutFunction;\n }\n | undefined;\n}\n\nexport const timeoutProvider: TimeoutProvider = {\n // When accessing the delegate, use the variable rather than `this` so that\n // the functions can be called without being bound to the provider.\n setTimeout(handler: () => void, timeout?: number, ...args) {\n const { delegate } = timeoutProvider;\n if (delegate?.setTimeout) {\n return delegate.setTimeout(handler, timeout, ...args);\n }\n return setTimeout(handler, timeout, ...args);\n },\n clearTimeout(handle) {\n const { delegate } = timeoutProvider;\n return (delegate?.clearTimeout || clearTimeout)(handle as any);\n },\n delegate: undefined,\n};\n", "import { config } from '../config';\nimport { timeoutProvider } from '../scheduler/timeoutProvider';\n\n/**\n * Handles an error on another job either with the user-configured {@link onUnhandledError},\n * or by throwing it on that new job so it can be picked up by `window.onerror`, `process.on('error')`, etc.\n *\n * This should be called whenever there is an error that is out-of-band with the subscription\n * or when an error hits a terminal boundary of the subscription and no error handler was provided.\n *\n * @param err the error to report\n */\nexport function reportUnhandledError(err: any) {\n timeoutProvider.setTimeout(() => {\n const { onUnhandledError } = config;\n if (onUnhandledError) {\n // Execute the user-configured error handler.\n onUnhandledError(err);\n } else {\n // Throw so it is picked up by the runtime's uncaught error mechanism.\n throw err;\n }\n });\n}\n", "/* tslint:disable:no-empty */\nexport function noop() { }\n", "import { CompleteNotification, NextNotification, ErrorNotification } from './types';\n\n/**\n * A completion object optimized for memory use and created to be the\n * same \"shape\" as other notifications in v8.\n * @internal\n */\nexport const COMPLETE_NOTIFICATION = (() => createNotification('C', undefined, undefined) as CompleteNotification)();\n\n/**\n * Internal use only. Creates an optimized error notification that is the same \"shape\"\n * as other notifications.\n * @internal\n */\nexport function errorNotification(error: any): ErrorNotification {\n return createNotification('E', undefined, error) as any;\n}\n\n/**\n * Internal use only. Creates an optimized next notification that is the same \"shape\"\n * as other notifications.\n * @internal\n */\nexport function nextNotification(value: T) {\n return createNotification('N', value, undefined) as NextNotification;\n}\n\n/**\n * Ensures that all notifications created internally have the same \"shape\" in v8.\n *\n * TODO: This is only exported to support a crazy legacy test in `groupBy`.\n * @internal\n */\nexport function createNotification(kind: 'N' | 'E' | 'C', value: any, error: any) {\n return {\n kind,\n value,\n error,\n };\n}\n", "import { config } from '../config';\n\nlet context: { errorThrown: boolean; error: any } | null = null;\n\n/**\n * Handles dealing with errors for super-gross mode. Creates a context, in which\n * any synchronously thrown errors will be passed to {@link captureError}. Which\n * will record the error such that it will be rethrown after the call back is complete.\n * TODO: Remove in v8\n * @param cb An immediately executed function.\n */\nexport function errorContext(cb: () => void) {\n if (config.useDeprecatedSynchronousErrorHandling) {\n const isRoot = !context;\n if (isRoot) {\n context = { errorThrown: false, error: null };\n }\n cb();\n if (isRoot) {\n const { errorThrown, error } = context!;\n context = null;\n if (errorThrown) {\n throw error;\n }\n }\n } else {\n // This is the general non-deprecated path for everyone that\n // isn't crazy enough to use super-gross mode (useDeprecatedSynchronousErrorHandling)\n cb();\n }\n}\n\n/**\n * Captures errors only in super-gross mode.\n * @param err the error to capture\n */\nexport function captureError(err: any) {\n if (config.useDeprecatedSynchronousErrorHandling && context) {\n context.errorThrown = true;\n context.error = err;\n }\n}\n", "import { isFunction } from './util/isFunction';\nimport { Observer, ObservableNotification } from './types';\nimport { isSubscription, Subscription } from './Subscription';\nimport { config } from './config';\nimport { reportUnhandledError } from './util/reportUnhandledError';\nimport { noop } from './util/noop';\nimport { nextNotification, errorNotification, COMPLETE_NOTIFICATION } from './NotificationFactories';\nimport { timeoutProvider } from './scheduler/timeoutProvider';\nimport { captureError } from './util/errorContext';\n\n/**\n * Implements the {@link Observer} interface and extends the\n * {@link Subscription} class. While the {@link Observer} is the public API for\n * consuming the values of an {@link Observable}, all Observers get converted to\n * a Subscriber, in order to provide Subscription-like capabilities such as\n * `unsubscribe`. Subscriber is a common type in RxJS, and crucial for\n * implementing operators, but it is rarely used as a public API.\n *\n * @class Subscriber\n */\nexport class Subscriber extends Subscription implements Observer {\n /**\n * A static factory for a Subscriber, given a (potentially partial) definition\n * of an Observer.\n * @param next The `next` callback of an Observer.\n * @param error The `error` callback of an\n * Observer.\n * @param complete The `complete` callback of an\n * Observer.\n * @return A Subscriber wrapping the (partially defined)\n * Observer represented by the given arguments.\n * @nocollapse\n * @deprecated Do not use. Will be removed in v8. There is no replacement for this\n * method, and there is no reason to be creating instances of `Subscriber` directly.\n * If you have a specific use case, please file an issue.\n */\n static create(next?: (x?: T) => void, error?: (e?: any) => void, complete?: () => void): Subscriber {\n return new SafeSubscriber(next, error, complete);\n }\n\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n protected isStopped: boolean = false;\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n protected destination: Subscriber | Observer; // this `any` is the escape hatch to erase extra type param (e.g. R)\n\n /**\n * @deprecated Internal implementation detail, do not use directly. Will be made internal in v8.\n * There is no reason to directly create an instance of Subscriber. This type is exported for typings reasons.\n */\n constructor(destination?: Subscriber | Observer) {\n super();\n if (destination) {\n this.destination = destination;\n // Automatically chain subscriptions together here.\n // if destination is a Subscription, then it is a Subscriber.\n if (isSubscription(destination)) {\n destination.add(this);\n }\n } else {\n this.destination = EMPTY_OBSERVER;\n }\n }\n\n /**\n * The {@link Observer} callback to receive notifications of type `next` from\n * the Observable, with a value. The Observable may call this method 0 or more\n * times.\n * @param {T} [value] The `next` value.\n * @return {void}\n */\n next(value?: T): void {\n if (this.isStopped) {\n handleStoppedNotification(nextNotification(value), this);\n } else {\n this._next(value!);\n }\n }\n\n /**\n * The {@link Observer} callback to receive notifications of type `error` from\n * the Observable, with an attached `Error`. Notifies the Observer that\n * the Observable has experienced an error condition.\n * @param {any} [err] The `error` exception.\n * @return {void}\n */\n error(err?: any): void {\n if (this.isStopped) {\n handleStoppedNotification(errorNotification(err), this);\n } else {\n this.isStopped = true;\n this._error(err);\n }\n }\n\n /**\n * The {@link Observer} callback to receive a valueless notification of type\n * `complete` from the Observable. Notifies the Observer that the Observable\n * has finished sending push-based notifications.\n * @return {void}\n */\n complete(): void {\n if (this.isStopped) {\n handleStoppedNotification(COMPLETE_NOTIFICATION, this);\n } else {\n this.isStopped = true;\n this._complete();\n }\n }\n\n unsubscribe(): void {\n if (!this.closed) {\n this.isStopped = true;\n super.unsubscribe();\n this.destination = null!;\n }\n }\n\n protected _next(value: T): void {\n this.destination.next(value);\n }\n\n protected _error(err: any): void {\n try {\n this.destination.error(err);\n } finally {\n this.unsubscribe();\n }\n }\n\n protected _complete(): void {\n try {\n this.destination.complete();\n } finally {\n this.unsubscribe();\n }\n }\n}\n\n/**\n * This bind is captured here because we want to be able to have\n * compatibility with monoid libraries that tend to use a method named\n * `bind`. In particular, a library called Monio requires this.\n */\nconst _bind = Function.prototype.bind;\n\nfunction bind any>(fn: Fn, thisArg: any): Fn {\n return _bind.call(fn, thisArg);\n}\n\n/**\n * Internal optimization only, DO NOT EXPOSE.\n * @internal\n */\nclass ConsumerObserver implements Observer {\n constructor(private partialObserver: Partial>) {}\n\n next(value: T): void {\n const { partialObserver } = this;\n if (partialObserver.next) {\n try {\n partialObserver.next(value);\n } catch (error) {\n handleUnhandledError(error);\n }\n }\n }\n\n error(err: any): void {\n const { partialObserver } = this;\n if (partialObserver.error) {\n try {\n partialObserver.error(err);\n } catch (error) {\n handleUnhandledError(error);\n }\n } else {\n handleUnhandledError(err);\n }\n }\n\n complete(): void {\n const { partialObserver } = this;\n if (partialObserver.complete) {\n try {\n partialObserver.complete();\n } catch (error) {\n handleUnhandledError(error);\n }\n }\n }\n}\n\nexport class SafeSubscriber extends Subscriber {\n constructor(\n observerOrNext?: Partial> | ((value: T) => void) | null,\n error?: ((e?: any) => void) | null,\n complete?: (() => void) | null\n ) {\n super();\n\n let partialObserver: Partial>;\n if (isFunction(observerOrNext) || !observerOrNext) {\n // The first argument is a function, not an observer. The next\n // two arguments *could* be observers, or they could be empty.\n partialObserver = {\n next: (observerOrNext ?? undefined) as (((value: T) => void) | undefined),\n error: error ?? undefined,\n complete: complete ?? undefined,\n };\n } else {\n // The first argument is a partial observer.\n let context: any;\n if (this && config.useDeprecatedNextContext) {\n // This is a deprecated path that made `this.unsubscribe()` available in\n // next handler functions passed to subscribe. This only exists behind a flag\n // now, as it is *very* slow.\n context = Object.create(observerOrNext);\n context.unsubscribe = () => this.unsubscribe();\n partialObserver = {\n next: observerOrNext.next && bind(observerOrNext.next, context),\n error: observerOrNext.error && bind(observerOrNext.error, context),\n complete: observerOrNext.complete && bind(observerOrNext.complete, context),\n };\n } else {\n // The \"normal\" path. Just use the partial observer directly.\n partialObserver = observerOrNext;\n }\n }\n\n // Wrap the partial observer to ensure it's a full observer, and\n // make sure proper error handling is accounted for.\n this.destination = new ConsumerObserver(partialObserver);\n }\n}\n\nfunction handleUnhandledError(error: any) {\n if (config.useDeprecatedSynchronousErrorHandling) {\n captureError(error);\n } else {\n // Ideal path, we report this as an unhandled error,\n // which is thrown on a new call stack.\n reportUnhandledError(error);\n }\n}\n\n/**\n * An error handler used when no error handler was supplied\n * to the SafeSubscriber -- meaning no error handler was supplied\n * do the `subscribe` call on our observable.\n * @param err The error to handle\n */\nfunction defaultErrorHandler(err: any) {\n throw err;\n}\n\n/**\n * A handler for notifications that cannot be sent to a stopped subscriber.\n * @param notification The notification being sent\n * @param subscriber The stopped subscriber\n */\nfunction handleStoppedNotification(notification: ObservableNotification, subscriber: Subscriber) {\n const { onStoppedNotification } = config;\n onStoppedNotification && timeoutProvider.setTimeout(() => onStoppedNotification(notification, subscriber));\n}\n\n/**\n * The observer used as a stub for subscriptions where the user did not\n * pass any arguments to `subscribe`. Comes with the default error handling\n * behavior.\n */\nexport const EMPTY_OBSERVER: Readonly> & { closed: true } = {\n closed: true,\n next: noop,\n error: defaultErrorHandler,\n complete: noop,\n};\n", "/**\n * Symbol.observable or a string \"@@observable\". Used for interop\n *\n * @deprecated We will no longer be exporting this symbol in upcoming versions of RxJS.\n * Instead polyfill and use Symbol.observable directly *or* use https://www.npmjs.com/package/symbol-observable\n */\nexport const observable: string | symbol = (() => (typeof Symbol === 'function' && Symbol.observable) || '@@observable')();\n", "/**\n * This function takes one parameter and just returns it. Simply put,\n * this is like `(x: T): T => x`.\n *\n * ## Examples\n *\n * This is useful in some cases when using things like `mergeMap`\n *\n * ```ts\n * import { interval, take, map, range, mergeMap, identity } from 'rxjs';\n *\n * const source$ = interval(1000).pipe(take(5));\n *\n * const result$ = source$.pipe(\n * map(i => range(i)),\n * mergeMap(identity) // same as mergeMap(x => x)\n * );\n *\n * result$.subscribe({\n * next: console.log\n * });\n * ```\n *\n * Or when you want to selectively apply an operator\n *\n * ```ts\n * import { interval, take, identity } from 'rxjs';\n *\n * const shouldLimit = () => Math.random() < 0.5;\n *\n * const source$ = interval(1000);\n *\n * const result$ = source$.pipe(shouldLimit() ? take(5) : identity);\n *\n * result$.subscribe({\n * next: console.log\n * });\n * ```\n *\n * @param x Any value that is returned by this function\n * @returns The value passed as the first parameter to this function\n */\nexport function identity(x: T): T {\n return x;\n}\n", "import { identity } from './identity';\nimport { UnaryFunction } from '../types';\n\nexport function pipe(): typeof identity;\nexport function pipe(fn1: UnaryFunction): UnaryFunction;\nexport function pipe(fn1: UnaryFunction, fn2: UnaryFunction): UnaryFunction;\nexport function pipe(fn1: UnaryFunction, fn2: UnaryFunction, fn3: UnaryFunction): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction,\n fn7: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction,\n fn7: UnaryFunction,\n fn8: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction,\n fn7: UnaryFunction,\n fn8: UnaryFunction,\n fn9: UnaryFunction\n): UnaryFunction;\nexport function pipe(\n fn1: UnaryFunction,\n fn2: UnaryFunction,\n fn3: UnaryFunction,\n fn4: UnaryFunction,\n fn5: UnaryFunction,\n fn6: UnaryFunction,\n fn7: UnaryFunction,\n fn8: UnaryFunction,\n fn9: UnaryFunction,\n ...fns: UnaryFunction[]\n): UnaryFunction;\n\n/**\n * pipe() can be called on one or more functions, each of which can take one argument (\"UnaryFunction\")\n * and uses it to return a value.\n * It returns a function that takes one argument, passes it to the first UnaryFunction, and then\n * passes the result to the next one, passes that result to the next one, and so on. \n */\nexport function pipe(...fns: Array>): UnaryFunction {\n return pipeFromArray(fns);\n}\n\n/** @internal */\nexport function pipeFromArray(fns: Array>): UnaryFunction {\n if (fns.length === 0) {\n return identity as UnaryFunction;\n }\n\n if (fns.length === 1) {\n return fns[0];\n }\n\n return function piped(input: T): R {\n return fns.reduce((prev: any, fn: UnaryFunction) => fn(prev), input as any);\n };\n}\n", "import { Operator } from './Operator';\nimport { SafeSubscriber, Subscriber } from './Subscriber';\nimport { isSubscription, Subscription } from './Subscription';\nimport { TeardownLogic, OperatorFunction, Subscribable, Observer } from './types';\nimport { observable as Symbol_observable } from './symbol/observable';\nimport { pipeFromArray } from './util/pipe';\nimport { config } from './config';\nimport { isFunction } from './util/isFunction';\nimport { errorContext } from './util/errorContext';\n\n/**\n * A representation of any set of values over any amount of time. This is the most basic building block\n * of RxJS.\n *\n * @class Observable\n */\nexport class Observable implements Subscribable {\n /**\n * @deprecated Internal implementation detail, do not use directly. Will be made internal in v8.\n */\n source: Observable | undefined;\n\n /**\n * @deprecated Internal implementation detail, do not use directly. Will be made internal in v8.\n */\n operator: Operator | undefined;\n\n /**\n * @constructor\n * @param {Function} subscribe the function that is called when the Observable is\n * initially subscribed to. This function is given a Subscriber, to which new values\n * can be `next`ed, or an `error` method can be called to raise an error, or\n * `complete` can be called to notify of a successful completion.\n */\n constructor(subscribe?: (this: Observable, subscriber: Subscriber) => TeardownLogic) {\n if (subscribe) {\n this._subscribe = subscribe;\n }\n }\n\n // HACK: Since TypeScript inherits static properties too, we have to\n // fight against TypeScript here so Subject can have a different static create signature\n /**\n * Creates a new Observable by calling the Observable constructor\n * @owner Observable\n * @method create\n * @param {Function} subscribe? the subscriber function to be passed to the Observable constructor\n * @return {Observable} a new observable\n * @nocollapse\n * @deprecated Use `new Observable()` instead. Will be removed in v8.\n */\n static create: (...args: any[]) => any = (subscribe?: (subscriber: Subscriber) => TeardownLogic) => {\n return new Observable(subscribe);\n };\n\n /**\n * Creates a new Observable, with this Observable instance as the source, and the passed\n * operator defined as the new observable's operator.\n * @method lift\n * @param operator the operator defining the operation to take on the observable\n * @return a new observable with the Operator applied\n * @deprecated Internal implementation detail, do not use directly. Will be made internal in v8.\n * If you have implemented an operator using `lift`, it is recommended that you create an\n * operator by simply returning `new Observable()` directly. See \"Creating new operators from\n * scratch\" section here: https://rxjs.dev/guide/operators\n */\n lift(operator?: Operator): Observable {\n const observable = new Observable();\n observable.source = this;\n observable.operator = operator;\n return observable;\n }\n\n subscribe(observerOrNext?: Partial> | ((value: T) => void)): Subscription;\n /** @deprecated Instead of passing separate callback arguments, use an observer argument. Signatures taking separate callback arguments will be removed in v8. Details: https://rxjs.dev/deprecations/subscribe-arguments */\n subscribe(next?: ((value: T) => void) | null, error?: ((error: any) => void) | null, complete?: (() => void) | null): Subscription;\n /**\n * Invokes an execution of an Observable and registers Observer handlers for notifications it will emit.\n *\n * Use it when you have all these Observables, but still nothing is happening.\n *\n * `subscribe` is not a regular operator, but a method that calls Observable's internal `subscribe` function. It\n * might be for example a function that you passed to Observable's constructor, but most of the time it is\n * a library implementation, which defines what will be emitted by an Observable, and when it be will emitted. This means\n * that calling `subscribe` is actually the moment when Observable starts its work, not when it is created, as it is often\n * the thought.\n *\n * Apart from starting the execution of an Observable, this method allows you to listen for values\n * that an Observable emits, as well as for when it completes or errors. You can achieve this in two\n * of the following ways.\n *\n * The first way is creating an object that implements {@link Observer} interface. It should have methods\n * defined by that interface, but note that it should be just a regular JavaScript object, which you can create\n * yourself in any way you want (ES6 class, classic function constructor, object literal etc.). In particular, do\n * not attempt to use any RxJS implementation details to create Observers - you don't need them. Remember also\n * that your object does not have to implement all methods. If you find yourself creating a method that doesn't\n * do anything, you can simply omit it. Note however, if the `error` method is not provided and an error happens,\n * it will be thrown asynchronously. Errors thrown asynchronously cannot be caught using `try`/`catch`. Instead,\n * use the {@link onUnhandledError} configuration option or use a runtime handler (like `window.onerror` or\n * `process.on('error)`) to be notified of unhandled errors. Because of this, it's recommended that you provide\n * an `error` method to avoid missing thrown errors.\n *\n * The second way is to give up on Observer object altogether and simply provide callback functions in place of its methods.\n * This means you can provide three functions as arguments to `subscribe`, where the first function is equivalent\n * of a `next` method, the second of an `error` method and the third of a `complete` method. Just as in case of an Observer,\n * if you do not need to listen for something, you can omit a function by passing `undefined` or `null`,\n * since `subscribe` recognizes these functions by where they were placed in function call. When it comes\n * to the `error` function, as with an Observer, if not provided, errors emitted by an Observable will be thrown asynchronously.\n *\n * You can, however, subscribe with no parameters at all. This may be the case where you're not interested in terminal events\n * and you also handled emissions internally by using operators (e.g. using `tap`).\n *\n * Whichever style of calling `subscribe` you use, in both cases it returns a Subscription object.\n * This object allows you to call `unsubscribe` on it, which in turn will stop the work that an Observable does and will clean\n * up all resources that an Observable used. Note that cancelling a subscription will not call `complete` callback\n * provided to `subscribe` function, which is reserved for a regular completion signal that comes from an Observable.\n *\n * Remember that callbacks provided to `subscribe` are not guaranteed to be called asynchronously.\n * It is an Observable itself that decides when these functions will be called. For example {@link of}\n * by default emits all its values synchronously. Always check documentation for how given Observable\n * will behave when subscribed and if its default behavior can be modified with a `scheduler`.\n *\n * #### Examples\n *\n * Subscribe with an {@link guide/observer Observer}\n *\n * ```ts\n * import { of } from 'rxjs';\n *\n * const sumObserver = {\n * sum: 0,\n * next(value) {\n * console.log('Adding: ' + value);\n * this.sum = this.sum + value;\n * },\n * error() {\n * // We actually could just remove this method,\n * // since we do not really care about errors right now.\n * },\n * complete() {\n * console.log('Sum equals: ' + this.sum);\n * }\n * };\n *\n * of(1, 2, 3) // Synchronously emits 1, 2, 3 and then completes.\n * .subscribe(sumObserver);\n *\n * // Logs:\n * // 'Adding: 1'\n * // 'Adding: 2'\n * // 'Adding: 3'\n * // 'Sum equals: 6'\n * ```\n *\n * Subscribe with functions ({@link deprecations/subscribe-arguments deprecated})\n *\n * ```ts\n * import { of } from 'rxjs'\n *\n * let sum = 0;\n *\n * of(1, 2, 3).subscribe(\n * value => {\n * console.log('Adding: ' + value);\n * sum = sum + value;\n * },\n * undefined,\n * () => console.log('Sum equals: ' + sum)\n * );\n *\n * // Logs:\n * // 'Adding: 1'\n * // 'Adding: 2'\n * // 'Adding: 3'\n * // 'Sum equals: 6'\n * ```\n *\n * Cancel a subscription\n *\n * ```ts\n * import { interval } from 'rxjs';\n *\n * const subscription = interval(1000).subscribe({\n * next(num) {\n * console.log(num)\n * },\n * complete() {\n * // Will not be called, even when cancelling subscription.\n * console.log('completed!');\n * }\n * });\n *\n * setTimeout(() => {\n * subscription.unsubscribe();\n * console.log('unsubscribed!');\n * }, 2500);\n *\n * // Logs:\n * // 0 after 1s\n * // 1 after 2s\n * // 'unsubscribed!' after 2.5s\n * ```\n *\n * @param {Observer|Function} observerOrNext (optional) Either an observer with methods to be called,\n * or the first of three possible handlers, which is the handler for each value emitted from the subscribed\n * Observable.\n * @param {Function} error (optional) A handler for a terminal event resulting from an error. If no error handler is provided,\n * the error will be thrown asynchronously as unhandled.\n * @param {Function} complete (optional) A handler for a terminal event resulting from successful completion.\n * @return {Subscription} a subscription reference to the registered handlers\n * @method subscribe\n */\n subscribe(\n observerOrNext?: Partial> | ((value: T) => void) | null,\n error?: ((error: any) => void) | null,\n complete?: (() => void) | null\n ): Subscription {\n const subscriber = isSubscriber(observerOrNext) ? observerOrNext : new SafeSubscriber(observerOrNext, error, complete);\n\n errorContext(() => {\n const { operator, source } = this;\n subscriber.add(\n operator\n ? // We're dealing with a subscription in the\n // operator chain to one of our lifted operators.\n operator.call(subscriber, source)\n : source\n ? // If `source` has a value, but `operator` does not, something that\n // had intimate knowledge of our API, like our `Subject`, must have\n // set it. We're going to just call `_subscribe` directly.\n this._subscribe(subscriber)\n : // In all other cases, we're likely wrapping a user-provided initializer\n // function, so we need to catch errors and handle them appropriately.\n this._trySubscribe(subscriber)\n );\n });\n\n return subscriber;\n }\n\n /** @internal */\n protected _trySubscribe(sink: Subscriber): TeardownLogic {\n try {\n return this._subscribe(sink);\n } catch (err) {\n // We don't need to return anything in this case,\n // because it's just going to try to `add()` to a subscription\n // above.\n sink.error(err);\n }\n }\n\n /**\n * Used as a NON-CANCELLABLE means of subscribing to an observable, for use with\n * APIs that expect promises, like `async/await`. You cannot unsubscribe from this.\n *\n * **WARNING**: Only use this with observables you *know* will complete. If the source\n * observable does not complete, you will end up with a promise that is hung up, and\n * potentially all of the state of an async function hanging out in memory. To avoid\n * this situation, look into adding something like {@link timeout}, {@link take},\n * {@link takeWhile}, or {@link takeUntil} amongst others.\n *\n * #### Example\n *\n * ```ts\n * import { interval, take } from 'rxjs';\n *\n * const source$ = interval(1000).pipe(take(4));\n *\n * async function getTotal() {\n * let total = 0;\n *\n * await source$.forEach(value => {\n * total += value;\n * console.log('observable -> ' + value);\n * });\n *\n * return total;\n * }\n *\n * getTotal().then(\n * total => console.log('Total: ' + total)\n * );\n *\n * // Expected:\n * // 'observable -> 0'\n * // 'observable -> 1'\n * // 'observable -> 2'\n * // 'observable -> 3'\n * // 'Total: 6'\n * ```\n *\n * @param next a handler for each value emitted by the observable\n * @return a promise that either resolves on observable completion or\n * rejects with the handled error\n */\n forEach(next: (value: T) => void): Promise;\n\n /**\n * @param next a handler for each value emitted by the observable\n * @param promiseCtor a constructor function used to instantiate the Promise\n * @return a promise that either resolves on observable completion or\n * rejects with the handled error\n * @deprecated Passing a Promise constructor will no longer be available\n * in upcoming versions of RxJS. This is because it adds weight to the library, for very\n * little benefit. If you need this functionality, it is recommended that you either\n * polyfill Promise, or you create an adapter to convert the returned native promise\n * to whatever promise implementation you wanted. Will be removed in v8.\n */\n forEach(next: (value: T) => void, promiseCtor: PromiseConstructorLike): Promise;\n\n forEach(next: (value: T) => void, promiseCtor?: PromiseConstructorLike): Promise {\n promiseCtor = getPromiseCtor(promiseCtor);\n\n return new promiseCtor((resolve, reject) => {\n const subscriber = new SafeSubscriber({\n next: (value) => {\n try {\n next(value);\n } catch (err) {\n reject(err);\n subscriber.unsubscribe();\n }\n },\n error: reject,\n complete: resolve,\n });\n this.subscribe(subscriber);\n }) as Promise;\n }\n\n /** @internal */\n protected _subscribe(subscriber: Subscriber): TeardownLogic {\n return this.source?.subscribe(subscriber);\n }\n\n /**\n * An interop point defined by the es7-observable spec https://github.com/zenparsing/es-observable\n * @method Symbol.observable\n * @return {Observable} this instance of the observable\n */\n [Symbol_observable]() {\n return this;\n }\n\n /* tslint:disable:max-line-length */\n pipe(): Observable;\n pipe(op1: OperatorFunction): Observable;\n pipe(op1: OperatorFunction, op2: OperatorFunction): Observable;\n pipe(op1: OperatorFunction, op2: OperatorFunction, op3: OperatorFunction): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction,\n op7: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction,\n op7: OperatorFunction,\n op8: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction,\n op7: OperatorFunction,\n op8: OperatorFunction,\n op9: OperatorFunction\n ): Observable;\n pipe(\n op1: OperatorFunction,\n op2: OperatorFunction,\n op3: OperatorFunction,\n op4: OperatorFunction,\n op5: OperatorFunction,\n op6: OperatorFunction,\n op7: OperatorFunction,\n op8: OperatorFunction,\n op9: OperatorFunction,\n ...operations: OperatorFunction[]\n ): Observable;\n /* tslint:enable:max-line-length */\n\n /**\n * Used to stitch together functional operators into a chain.\n * @method pipe\n * @return {Observable} the Observable result of all of the operators having\n * been called in the order they were passed in.\n *\n * ## Example\n *\n * ```ts\n * import { interval, filter, map, scan } from 'rxjs';\n *\n * interval(1000)\n * .pipe(\n * filter(x => x % 2 === 0),\n * map(x => x + x),\n * scan((acc, x) => acc + x)\n * )\n * .subscribe(x => console.log(x));\n * ```\n */\n pipe(...operations: OperatorFunction[]): Observable {\n return pipeFromArray(operations)(this);\n }\n\n /* tslint:disable:max-line-length */\n /** @deprecated Replaced with {@link firstValueFrom} and {@link lastValueFrom}. Will be removed in v8. Details: https://rxjs.dev/deprecations/to-promise */\n toPromise(): Promise;\n /** @deprecated Replaced with {@link firstValueFrom} and {@link lastValueFrom}. Will be removed in v8. Details: https://rxjs.dev/deprecations/to-promise */\n toPromise(PromiseCtor: typeof Promise): Promise;\n /** @deprecated Replaced with {@link firstValueFrom} and {@link lastValueFrom}. Will be removed in v8. Details: https://rxjs.dev/deprecations/to-promise */\n toPromise(PromiseCtor: PromiseConstructorLike): Promise;\n /* tslint:enable:max-line-length */\n\n /**\n * Subscribe to this Observable and get a Promise resolving on\n * `complete` with the last emission (if any).\n *\n * **WARNING**: Only use this with observables you *know* will complete. If the source\n * observable does not complete, you will end up with a promise that is hung up, and\n * potentially all of the state of an async function hanging out in memory. To avoid\n * this situation, look into adding something like {@link timeout}, {@link take},\n * {@link takeWhile}, or {@link takeUntil} amongst others.\n *\n * @method toPromise\n * @param [promiseCtor] a constructor function used to instantiate\n * the Promise\n * @return A Promise that resolves with the last value emit, or\n * rejects on an error. If there were no emissions, Promise\n * resolves with undefined.\n * @deprecated Replaced with {@link firstValueFrom} and {@link lastValueFrom}. Will be removed in v8. Details: https://rxjs.dev/deprecations/to-promise\n */\n toPromise(promiseCtor?: PromiseConstructorLike): Promise {\n promiseCtor = getPromiseCtor(promiseCtor);\n\n return new promiseCtor((resolve, reject) => {\n let value: T | undefined;\n this.subscribe(\n (x: T) => (value = x),\n (err: any) => reject(err),\n () => resolve(value)\n );\n }) as Promise;\n }\n}\n\n/**\n * Decides between a passed promise constructor from consuming code,\n * A default configured promise constructor, and the native promise\n * constructor and returns it. If nothing can be found, it will throw\n * an error.\n * @param promiseCtor The optional promise constructor to passed by consuming code\n */\nfunction getPromiseCtor(promiseCtor: PromiseConstructorLike | undefined) {\n return promiseCtor ?? config.Promise ?? Promise;\n}\n\nfunction isObserver(value: any): value is Observer {\n return value && isFunction(value.next) && isFunction(value.error) && isFunction(value.complete);\n}\n\nfunction isSubscriber(value: any): value is Subscriber {\n return (value && value instanceof Subscriber) || (isObserver(value) && isSubscription(value));\n}\n", "import { Observable } from '../Observable';\nimport { Subscriber } from '../Subscriber';\nimport { OperatorFunction } from '../types';\nimport { isFunction } from './isFunction';\n\n/**\n * Used to determine if an object is an Observable with a lift function.\n */\nexport function hasLift(source: any): source is { lift: InstanceType['lift'] } {\n return isFunction(source?.lift);\n}\n\n/**\n * Creates an `OperatorFunction`. Used to define operators throughout the library in a concise way.\n * @param init The logic to connect the liftedSource to the subscriber at the moment of subscription.\n */\nexport function operate(\n init: (liftedSource: Observable, subscriber: Subscriber) => (() => void) | void\n): OperatorFunction {\n return (source: Observable) => {\n if (hasLift(source)) {\n return source.lift(function (this: Subscriber, liftedSource: Observable) {\n try {\n return init(liftedSource, this);\n } catch (err) {\n this.error(err);\n }\n });\n }\n throw new TypeError('Unable to lift unknown Observable type');\n };\n}\n", "import { Subscriber } from '../Subscriber';\n\n/**\n * Creates an instance of an `OperatorSubscriber`.\n * @param destination The downstream subscriber.\n * @param onNext Handles next values, only called if this subscriber is not stopped or closed. Any\n * error that occurs in this function is caught and sent to the `error` method of this subscriber.\n * @param onError Handles errors from the subscription, any errors that occur in this handler are caught\n * and send to the `destination` error handler.\n * @param onComplete Handles completion notification from the subscription. Any errors that occur in\n * this handler are sent to the `destination` error handler.\n * @param onFinalize Additional teardown logic here. This will only be called on teardown if the\n * subscriber itself is not already closed. This is called after all other teardown logic is executed.\n */\nexport function createOperatorSubscriber(\n destination: Subscriber,\n onNext?: (value: T) => void,\n onComplete?: () => void,\n onError?: (err: any) => void,\n onFinalize?: () => void\n): Subscriber {\n return new OperatorSubscriber(destination, onNext, onComplete, onError, onFinalize);\n}\n\n/**\n * A generic helper for allowing operators to be created with a Subscriber and\n * use closures to capture necessary state from the operator function itself.\n */\nexport class OperatorSubscriber extends Subscriber {\n /**\n * Creates an instance of an `OperatorSubscriber`.\n * @param destination The downstream subscriber.\n * @param onNext Handles next values, only called if this subscriber is not stopped or closed. Any\n * error that occurs in this function is caught and sent to the `error` method of this subscriber.\n * @param onError Handles errors from the subscription, any errors that occur in this handler are caught\n * and send to the `destination` error handler.\n * @param onComplete Handles completion notification from the subscription. Any errors that occur in\n * this handler are sent to the `destination` error handler.\n * @param onFinalize Additional finalization logic here. This will only be called on finalization if the\n * subscriber itself is not already closed. This is called after all other finalization logic is executed.\n * @param shouldUnsubscribe An optional check to see if an unsubscribe call should truly unsubscribe.\n * NOTE: This currently **ONLY** exists to support the strange behavior of {@link groupBy}, where unsubscription\n * to the resulting observable does not actually disconnect from the source if there are active subscriptions\n * to any grouped observable. (DO NOT EXPOSE OR USE EXTERNALLY!!!)\n */\n constructor(\n destination: Subscriber,\n onNext?: (value: T) => void,\n onComplete?: () => void,\n onError?: (err: any) => void,\n private onFinalize?: () => void,\n private shouldUnsubscribe?: () => boolean\n ) {\n // It's important - for performance reasons - that all of this class's\n // members are initialized and that they are always initialized in the same\n // order. This will ensure that all OperatorSubscriber instances have the\n // same hidden class in V8. This, in turn, will help keep the number of\n // hidden classes involved in property accesses within the base class as\n // low as possible. If the number of hidden classes involved exceeds four,\n // the property accesses will become megamorphic and performance penalties\n // will be incurred - i.e. inline caches won't be used.\n //\n // The reasons for ensuring all instances have the same hidden class are\n // further discussed in this blog post from Benedikt Meurer:\n // https://benediktmeurer.de/2018/03/23/impact-of-polymorphism-on-component-based-frameworks-like-react/\n super(destination);\n this._next = onNext\n ? function (this: OperatorSubscriber, value: T) {\n try {\n onNext(value);\n } catch (err) {\n destination.error(err);\n }\n }\n : super._next;\n this._error = onError\n ? function (this: OperatorSubscriber, err: any) {\n try {\n onError(err);\n } catch (err) {\n // Send any errors that occur down stream.\n destination.error(err);\n } finally {\n // Ensure finalization.\n this.unsubscribe();\n }\n }\n : super._error;\n this._complete = onComplete\n ? function (this: OperatorSubscriber) {\n try {\n onComplete();\n } catch (err) {\n // Send any errors that occur down stream.\n destination.error(err);\n } finally {\n // Ensure finalization.\n this.unsubscribe();\n }\n }\n : super._complete;\n }\n\n unsubscribe() {\n if (!this.shouldUnsubscribe || this.shouldUnsubscribe()) {\n const { closed } = this;\n super.unsubscribe();\n // Execute additional teardown if we have any and we didn't already do so.\n !closed && this.onFinalize?.();\n }\n }\n}\n", "import { Subscription } from '../Subscription';\n\ninterface AnimationFrameProvider {\n schedule(callback: FrameRequestCallback): Subscription;\n requestAnimationFrame: typeof requestAnimationFrame;\n cancelAnimationFrame: typeof cancelAnimationFrame;\n delegate:\n | {\n requestAnimationFrame: typeof requestAnimationFrame;\n cancelAnimationFrame: typeof cancelAnimationFrame;\n }\n | undefined;\n}\n\nexport const animationFrameProvider: AnimationFrameProvider = {\n // When accessing the delegate, use the variable rather than `this` so that\n // the functions can be called without being bound to the provider.\n schedule(callback) {\n let request = requestAnimationFrame;\n let cancel: typeof cancelAnimationFrame | undefined = cancelAnimationFrame;\n const { delegate } = animationFrameProvider;\n if (delegate) {\n request = delegate.requestAnimationFrame;\n cancel = delegate.cancelAnimationFrame;\n }\n const handle = request((timestamp) => {\n // Clear the cancel function. The request has been fulfilled, so\n // attempting to cancel the request upon unsubscription would be\n // pointless.\n cancel = undefined;\n callback(timestamp);\n });\n return new Subscription(() => cancel?.(handle));\n },\n requestAnimationFrame(...args) {\n const { delegate } = animationFrameProvider;\n return (delegate?.requestAnimationFrame || requestAnimationFrame)(...args);\n },\n cancelAnimationFrame(...args) {\n const { delegate } = animationFrameProvider;\n return (delegate?.cancelAnimationFrame || cancelAnimationFrame)(...args);\n },\n delegate: undefined,\n};\n", "import { createErrorClass } from './createErrorClass';\n\nexport interface ObjectUnsubscribedError extends Error {}\n\nexport interface ObjectUnsubscribedErrorCtor {\n /**\n * @deprecated Internal implementation detail. Do not construct error instances.\n * Cannot be tagged as internal: https://github.com/ReactiveX/rxjs/issues/6269\n */\n new (): ObjectUnsubscribedError;\n}\n\n/**\n * An error thrown when an action is invalid because the object has been\n * unsubscribed.\n *\n * @see {@link Subject}\n * @see {@link BehaviorSubject}\n *\n * @class ObjectUnsubscribedError\n */\nexport const ObjectUnsubscribedError: ObjectUnsubscribedErrorCtor = createErrorClass(\n (_super) =>\n function ObjectUnsubscribedErrorImpl(this: any) {\n _super(this);\n this.name = 'ObjectUnsubscribedError';\n this.message = 'object unsubscribed';\n }\n);\n", "import { Operator } from './Operator';\nimport { Observable } from './Observable';\nimport { Subscriber } from './Subscriber';\nimport { Subscription, EMPTY_SUBSCRIPTION } from './Subscription';\nimport { Observer, SubscriptionLike, TeardownLogic } from './types';\nimport { ObjectUnsubscribedError } from './util/ObjectUnsubscribedError';\nimport { arrRemove } from './util/arrRemove';\nimport { errorContext } from './util/errorContext';\n\n/**\n * A Subject is a special type of Observable that allows values to be\n * multicasted to many Observers. Subjects are like EventEmitters.\n *\n * Every Subject is an Observable and an Observer. You can subscribe to a\n * Subject, and you can call next to feed values as well as error and complete.\n */\nexport class Subject extends Observable implements SubscriptionLike {\n closed = false;\n\n private currentObservers: Observer[] | null = null;\n\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n observers: Observer[] = [];\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n isStopped = false;\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n hasError = false;\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n thrownError: any = null;\n\n /**\n * Creates a \"subject\" by basically gluing an observer to an observable.\n *\n * @nocollapse\n * @deprecated Recommended you do not use. Will be removed at some point in the future. Plans for replacement still under discussion.\n */\n static create: (...args: any[]) => any = (destination: Observer, source: Observable): AnonymousSubject => {\n return new AnonymousSubject(destination, source);\n };\n\n constructor() {\n // NOTE: This must be here to obscure Observable's constructor.\n super();\n }\n\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n lift(operator: Operator): Observable {\n const subject = new AnonymousSubject(this, this);\n subject.operator = operator as any;\n return subject as any;\n }\n\n /** @internal */\n protected _throwIfClosed() {\n if (this.closed) {\n throw new ObjectUnsubscribedError();\n }\n }\n\n next(value: T) {\n errorContext(() => {\n this._throwIfClosed();\n if (!this.isStopped) {\n if (!this.currentObservers) {\n this.currentObservers = Array.from(this.observers);\n }\n for (const observer of this.currentObservers) {\n observer.next(value);\n }\n }\n });\n }\n\n error(err: any) {\n errorContext(() => {\n this._throwIfClosed();\n if (!this.isStopped) {\n this.hasError = this.isStopped = true;\n this.thrownError = err;\n const { observers } = this;\n while (observers.length) {\n observers.shift()!.error(err);\n }\n }\n });\n }\n\n complete() {\n errorContext(() => {\n this._throwIfClosed();\n if (!this.isStopped) {\n this.isStopped = true;\n const { observers } = this;\n while (observers.length) {\n observers.shift()!.complete();\n }\n }\n });\n }\n\n unsubscribe() {\n this.isStopped = this.closed = true;\n this.observers = this.currentObservers = null!;\n }\n\n get observed() {\n return this.observers?.length > 0;\n }\n\n /** @internal */\n protected _trySubscribe(subscriber: Subscriber): TeardownLogic {\n this._throwIfClosed();\n return super._trySubscribe(subscriber);\n }\n\n /** @internal */\n protected _subscribe(subscriber: Subscriber): Subscription {\n this._throwIfClosed();\n this._checkFinalizedStatuses(subscriber);\n return this._innerSubscribe(subscriber);\n }\n\n /** @internal */\n protected _innerSubscribe(subscriber: Subscriber) {\n const { hasError, isStopped, observers } = this;\n if (hasError || isStopped) {\n return EMPTY_SUBSCRIPTION;\n }\n this.currentObservers = null;\n observers.push(subscriber);\n return new Subscription(() => {\n this.currentObservers = null;\n arrRemove(observers, subscriber);\n });\n }\n\n /** @internal */\n protected _checkFinalizedStatuses(subscriber: Subscriber) {\n const { hasError, thrownError, isStopped } = this;\n if (hasError) {\n subscriber.error(thrownError);\n } else if (isStopped) {\n subscriber.complete();\n }\n }\n\n /**\n * Creates a new Observable with this Subject as the source. You can do this\n * to create custom Observer-side logic of the Subject and conceal it from\n * code that uses the Observable.\n * @return {Observable} Observable that the Subject casts to\n */\n asObservable(): Observable {\n const observable: any = new Observable();\n observable.source = this;\n return observable;\n }\n}\n\n/**\n * @class AnonymousSubject\n */\nexport class AnonymousSubject extends Subject {\n constructor(\n /** @deprecated Internal implementation detail, do not use directly. Will be made internal in v8. */\n public destination?: Observer,\n source?: Observable\n ) {\n super();\n this.source = source;\n }\n\n next(value: T) {\n this.destination?.next?.(value);\n }\n\n error(err: any) {\n this.destination?.error?.(err);\n }\n\n complete() {\n this.destination?.complete?.();\n }\n\n /** @internal */\n protected _subscribe(subscriber: Subscriber): Subscription {\n return this.source?.subscribe(subscriber) ?? EMPTY_SUBSCRIPTION;\n }\n}\n", "import { Subject } from './Subject';\nimport { Subscriber } from './Subscriber';\nimport { Subscription } from './Subscription';\n\n/**\n * A variant of Subject that requires an initial value and emits its current\n * value whenever it is subscribed to.\n *\n * @class BehaviorSubject\n */\nexport class BehaviorSubject extends Subject {\n constructor(private _value: T) {\n super();\n }\n\n get value(): T {\n return this.getValue();\n }\n\n /** @internal */\n protected _subscribe(subscriber: Subscriber): Subscription {\n const subscription = super._subscribe(subscriber);\n !subscription.closed && subscriber.next(this._value);\n return subscription;\n }\n\n getValue(): T {\n const { hasError, thrownError, _value } = this;\n if (hasError) {\n throw thrownError;\n }\n this._throwIfClosed();\n return _value;\n }\n\n next(value: T): void {\n super.next((this._value = value));\n }\n}\n", "import { TimestampProvider } from '../types';\n\ninterface DateTimestampProvider extends TimestampProvider {\n delegate: TimestampProvider | undefined;\n}\n\nexport const dateTimestampProvider: DateTimestampProvider = {\n now() {\n // Use the variable rather than `this` so that the function can be called\n // without being bound to the provider.\n return (dateTimestampProvider.delegate || Date).now();\n },\n delegate: undefined,\n};\n", "import { Subject } from './Subject';\nimport { TimestampProvider } from './types';\nimport { Subscriber } from './Subscriber';\nimport { Subscription } from './Subscription';\nimport { dateTimestampProvider } from './scheduler/dateTimestampProvider';\n\n/**\n * A variant of {@link Subject} that \"replays\" old values to new subscribers by emitting them when they first subscribe.\n *\n * `ReplaySubject` has an internal buffer that will store a specified number of values that it has observed. Like `Subject`,\n * `ReplaySubject` \"observes\" values by having them passed to its `next` method. When it observes a value, it will store that\n * value for a time determined by the configuration of the `ReplaySubject`, as passed to its constructor.\n *\n * When a new subscriber subscribes to the `ReplaySubject` instance, it will synchronously emit all values in its buffer in\n * a First-In-First-Out (FIFO) manner. The `ReplaySubject` will also complete, if it has observed completion; and it will\n * error if it has observed an error.\n *\n * There are two main configuration items to be concerned with:\n *\n * 1. `bufferSize` - This will determine how many items are stored in the buffer, defaults to infinite.\n * 2. `windowTime` - The amount of time to hold a value in the buffer before removing it from the buffer.\n *\n * Both configurations may exist simultaneously. So if you would like to buffer a maximum of 3 values, as long as the values\n * are less than 2 seconds old, you could do so with a `new ReplaySubject(3, 2000)`.\n *\n * ### Differences with BehaviorSubject\n *\n * `BehaviorSubject` is similar to `new ReplaySubject(1)`, with a couple of exceptions:\n *\n * 1. `BehaviorSubject` comes \"primed\" with a single value upon construction.\n * 2. `ReplaySubject` will replay values, even after observing an error, where `BehaviorSubject` will not.\n *\n * @see {@link Subject}\n * @see {@link BehaviorSubject}\n * @see {@link shareReplay}\n */\nexport class ReplaySubject extends Subject {\n private _buffer: (T | number)[] = [];\n private _infiniteTimeWindow = true;\n\n /**\n * @param bufferSize The size of the buffer to replay on subscription\n * @param windowTime The amount of time the buffered items will stay buffered\n * @param timestampProvider An object with a `now()` method that provides the current timestamp. This is used to\n * calculate the amount of time something has been buffered.\n */\n constructor(\n private _bufferSize = Infinity,\n private _windowTime = Infinity,\n private _timestampProvider: TimestampProvider = dateTimestampProvider\n ) {\n super();\n this._infiniteTimeWindow = _windowTime === Infinity;\n this._bufferSize = Math.max(1, _bufferSize);\n this._windowTime = Math.max(1, _windowTime);\n }\n\n next(value: T): void {\n const { isStopped, _buffer, _infiniteTimeWindow, _timestampProvider, _windowTime } = this;\n if (!isStopped) {\n _buffer.push(value);\n !_infiniteTimeWindow && _buffer.push(_timestampProvider.now() + _windowTime);\n }\n this._trimBuffer();\n super.next(value);\n }\n\n /** @internal */\n protected _subscribe(subscriber: Subscriber): Subscription {\n this._throwIfClosed();\n this._trimBuffer();\n\n const subscription = this._innerSubscribe(subscriber);\n\n const { _infiniteTimeWindow, _buffer } = this;\n // We use a copy here, so reentrant code does not mutate our array while we're\n // emitting it to a new subscriber.\n const copy = _buffer.slice();\n for (let i = 0; i < copy.length && !subscriber.closed; i += _infiniteTimeWindow ? 1 : 2) {\n subscriber.next(copy[i] as T);\n }\n\n this._checkFinalizedStatuses(subscriber);\n\n return subscription;\n }\n\n private _trimBuffer() {\n const { _bufferSize, _timestampProvider, _buffer, _infiniteTimeWindow } = this;\n // If we don't have an infinite buffer size, and we're over the length,\n // use splice to truncate the old buffer values off. Note that we have to\n // double the size for instances where we're not using an infinite time window\n // because we're storing the values and the timestamps in the same array.\n const adjustedBufferSize = (_infiniteTimeWindow ? 1 : 2) * _bufferSize;\n _bufferSize < Infinity && adjustedBufferSize < _buffer.length && _buffer.splice(0, _buffer.length - adjustedBufferSize);\n\n // Now, if we're not in an infinite time window, remove all values where the time is\n // older than what is allowed.\n if (!_infiniteTimeWindow) {\n const now = _timestampProvider.now();\n let last = 0;\n // Search the array for the first timestamp that isn't expired and\n // truncate the buffer up to that point.\n for (let i = 1; i < _buffer.length && (_buffer[i] as number) <= now; i += 2) {\n last = i;\n }\n last && _buffer.splice(0, last + 1);\n }\n }\n}\n", "import { Scheduler } from '../Scheduler';\nimport { Subscription } from '../Subscription';\nimport { SchedulerAction } from '../types';\n\n/**\n * A unit of work to be executed in a `scheduler`. An action is typically\n * created from within a {@link SchedulerLike} and an RxJS user does not need to concern\n * themselves about creating and manipulating an Action.\n *\n * ```ts\n * class Action extends Subscription {\n * new (scheduler: Scheduler, work: (state?: T) => void);\n * schedule(state?: T, delay: number = 0): Subscription;\n * }\n * ```\n *\n * @class Action\n */\nexport class Action extends Subscription {\n constructor(scheduler: Scheduler, work: (this: SchedulerAction, state?: T) => void) {\n super();\n }\n /**\n * Schedules this action on its parent {@link SchedulerLike} for execution. May be passed\n * some context object, `state`. May happen at some point in the future,\n * according to the `delay` parameter, if specified.\n * @param {T} [state] Some contextual data that the `work` function uses when\n * called by the Scheduler.\n * @param {number} [delay] Time to wait before executing the work, where the\n * time unit is implicit and defined by the Scheduler.\n * @return {void}\n */\n public schedule(state?: T, delay: number = 0): Subscription {\n return this;\n }\n}\n", "import type { TimerHandle } from './timerHandle';\ntype SetIntervalFunction = (handler: () => void, timeout?: number, ...args: any[]) => TimerHandle;\ntype ClearIntervalFunction = (handle: TimerHandle) => void;\n\ninterface IntervalProvider {\n setInterval: SetIntervalFunction;\n clearInterval: ClearIntervalFunction;\n delegate:\n | {\n setInterval: SetIntervalFunction;\n clearInterval: ClearIntervalFunction;\n }\n | undefined;\n}\n\nexport const intervalProvider: IntervalProvider = {\n // When accessing the delegate, use the variable rather than `this` so that\n // the functions can be called without being bound to the provider.\n setInterval(handler: () => void, timeout?: number, ...args) {\n const { delegate } = intervalProvider;\n if (delegate?.setInterval) {\n return delegate.setInterval(handler, timeout, ...args);\n }\n return setInterval(handler, timeout, ...args);\n },\n clearInterval(handle) {\n const { delegate } = intervalProvider;\n return (delegate?.clearInterval || clearInterval)(handle as any);\n },\n delegate: undefined,\n};\n", "import { Action } from './Action';\nimport { SchedulerAction } from '../types';\nimport { Subscription } from '../Subscription';\nimport { AsyncScheduler } from './AsyncScheduler';\nimport { intervalProvider } from './intervalProvider';\nimport { arrRemove } from '../util/arrRemove';\nimport { TimerHandle } from './timerHandle';\n\nexport class AsyncAction extends Action {\n public id: TimerHandle | undefined;\n public state?: T;\n // @ts-ignore: Property has no initializer and is not definitely assigned\n public delay: number;\n protected pending: boolean = false;\n\n constructor(protected scheduler: AsyncScheduler, protected work: (this: SchedulerAction, state?: T) => void) {\n super(scheduler, work);\n }\n\n public schedule(state?: T, delay: number = 0): Subscription {\n if (this.closed) {\n return this;\n }\n\n // Always replace the current state with the new state.\n this.state = state;\n\n const id = this.id;\n const scheduler = this.scheduler;\n\n //\n // Important implementation note:\n //\n // Actions only execute once by default, unless rescheduled from within the\n // scheduled callback. This allows us to implement single and repeat\n // actions via the same code path, without adding API surface area, as well\n // as mimic traditional recursion but across asynchronous boundaries.\n //\n // However, JS runtimes and timers distinguish between intervals achieved by\n // serial `setTimeout` calls vs. a single `setInterval` call. An interval of\n // serial `setTimeout` calls can be individually delayed, which delays\n // scheduling the next `setTimeout`, and so on. `setInterval` attempts to\n // guarantee the interval callback will be invoked more precisely to the\n // interval period, regardless of load.\n //\n // Therefore, we use `setInterval` to schedule single and repeat actions.\n // If the action reschedules itself with the same delay, the interval is not\n // canceled. If the action doesn't reschedule, or reschedules with a\n // different delay, the interval will be canceled after scheduled callback\n // execution.\n //\n if (id != null) {\n this.id = this.recycleAsyncId(scheduler, id, delay);\n }\n\n // Set the pending flag indicating that this action has been scheduled, or\n // has recursively rescheduled itself.\n this.pending = true;\n\n this.delay = delay;\n // If this action has already an async Id, don't request a new one.\n this.id = this.id ?? this.requestAsyncId(scheduler, this.id, delay);\n\n return this;\n }\n\n protected requestAsyncId(scheduler: AsyncScheduler, _id?: TimerHandle, delay: number = 0): TimerHandle {\n return intervalProvider.setInterval(scheduler.flush.bind(scheduler, this), delay);\n }\n\n protected recycleAsyncId(_scheduler: AsyncScheduler, id?: TimerHandle, delay: number | null = 0): TimerHandle | undefined {\n // If this action is rescheduled with the same delay time, don't clear the interval id.\n if (delay != null && this.delay === delay && this.pending === false) {\n return id;\n }\n // Otherwise, if the action's delay time is different from the current delay,\n // or the action has been rescheduled before it's executed, clear the interval id\n if (id != null) {\n intervalProvider.clearInterval(id);\n }\n\n return undefined;\n }\n\n /**\n * Immediately executes this action and the `work` it contains.\n * @return {any}\n */\n public execute(state: T, delay: number): any {\n if (this.closed) {\n return new Error('executing a cancelled action');\n }\n\n this.pending = false;\n const error = this._execute(state, delay);\n if (error) {\n return error;\n } else if (this.pending === false && this.id != null) {\n // Dequeue if the action didn't reschedule itself. Don't call\n // unsubscribe(), because the action could reschedule later.\n // For example:\n // ```\n // scheduler.schedule(function doWork(counter) {\n // /* ... I'm a busy worker bee ... */\n // var originalAction = this;\n // /* wait 100ms before rescheduling the action */\n // setTimeout(function () {\n // originalAction.schedule(counter + 1);\n // }, 100);\n // }, 1000);\n // ```\n this.id = this.recycleAsyncId(this.scheduler, this.id, null);\n }\n }\n\n protected _execute(state: T, _delay: number): any {\n let errored: boolean = false;\n let errorValue: any;\n try {\n this.work(state);\n } catch (e) {\n errored = true;\n // HACK: Since code elsewhere is relying on the \"truthiness\" of the\n // return here, we can't have it return \"\" or 0 or false.\n // TODO: Clean this up when we refactor schedulers mid-version-8 or so.\n errorValue = e ? e : new Error('Scheduled action threw falsy error');\n }\n if (errored) {\n this.unsubscribe();\n return errorValue;\n }\n }\n\n unsubscribe() {\n if (!this.closed) {\n const { id, scheduler } = this;\n const { actions } = scheduler;\n\n this.work = this.state = this.scheduler = null!;\n this.pending = false;\n\n arrRemove(actions, this);\n if (id != null) {\n this.id = this.recycleAsyncId(scheduler, id, null);\n }\n\n this.delay = null!;\n super.unsubscribe();\n }\n }\n}\n", "import { Action } from './scheduler/Action';\nimport { Subscription } from './Subscription';\nimport { SchedulerLike, SchedulerAction } from './types';\nimport { dateTimestampProvider } from './scheduler/dateTimestampProvider';\n\n/**\n * An execution context and a data structure to order tasks and schedule their\n * execution. Provides a notion of (potentially virtual) time, through the\n * `now()` getter method.\n *\n * Each unit of work in a Scheduler is called an `Action`.\n *\n * ```ts\n * class Scheduler {\n * now(): number;\n * schedule(work, delay?, state?): Subscription;\n * }\n * ```\n *\n * @class Scheduler\n * @deprecated Scheduler is an internal implementation detail of RxJS, and\n * should not be used directly. Rather, create your own class and implement\n * {@link SchedulerLike}. Will be made internal in v8.\n */\nexport class Scheduler implements SchedulerLike {\n public static now: () => number = dateTimestampProvider.now;\n\n constructor(private schedulerActionCtor: typeof Action, now: () => number = Scheduler.now) {\n this.now = now;\n }\n\n /**\n * A getter method that returns a number representing the current time\n * (at the time this function was called) according to the scheduler's own\n * internal clock.\n * @return {number} A number that represents the current time. May or may not\n * have a relation to wall-clock time. May or may not refer to a time unit\n * (e.g. milliseconds).\n */\n public now: () => number;\n\n /**\n * Schedules a function, `work`, for execution. May happen at some point in\n * the future, according to the `delay` parameter, if specified. May be passed\n * some context object, `state`, which will be passed to the `work` function.\n *\n * The given arguments will be processed an stored as an Action object in a\n * queue of actions.\n *\n * @param {function(state: ?T): ?Subscription} work A function representing a\n * task, or some unit of work to be executed by the Scheduler.\n * @param {number} [delay] Time to wait before executing the work, where the\n * time unit is implicit and defined by the Scheduler itself.\n * @param {T} [state] Some contextual data that the `work` function uses when\n * called by the Scheduler.\n * @return {Subscription} A subscription in order to be able to unsubscribe\n * the scheduled work.\n */\n public schedule(work: (this: SchedulerAction, state?: T) => void, delay: number = 0, state?: T): Subscription {\n return new this.schedulerActionCtor(this, work).schedule(state, delay);\n }\n}\n", "import { Scheduler } from '../Scheduler';\nimport { Action } from './Action';\nimport { AsyncAction } from './AsyncAction';\nimport { TimerHandle } from './timerHandle';\n\nexport class AsyncScheduler extends Scheduler {\n public actions: Array> = [];\n /**\n * A flag to indicate whether the Scheduler is currently executing a batch of\n * queued actions.\n * @type {boolean}\n * @internal\n */\n public _active: boolean = false;\n /**\n * An internal ID used to track the latest asynchronous task such as those\n * coming from `setTimeout`, `setInterval`, `requestAnimationFrame`, and\n * others.\n * @type {any}\n * @internal\n */\n public _scheduled: TimerHandle | undefined;\n\n constructor(SchedulerAction: typeof Action, now: () => number = Scheduler.now) {\n super(SchedulerAction, now);\n }\n\n public flush(action: AsyncAction): void {\n const { actions } = this;\n\n if (this._active) {\n actions.push(action);\n return;\n }\n\n let error: any;\n this._active = true;\n\n do {\n if ((error = action.execute(action.state, action.delay))) {\n break;\n }\n } while ((action = actions.shift()!)); // exhaust the scheduler queue\n\n this._active = false;\n\n if (error) {\n while ((action = actions.shift()!)) {\n action.unsubscribe();\n }\n throw error;\n }\n }\n}\n", "import { AsyncAction } from './AsyncAction';\nimport { AsyncScheduler } from './AsyncScheduler';\n\n/**\n *\n * Async Scheduler\n *\n * Schedule task as if you used setTimeout(task, duration)\n *\n * `async` scheduler schedules tasks asynchronously, by putting them on the JavaScript\n * event loop queue. It is best used to delay tasks in time or to schedule tasks repeating\n * in intervals.\n *\n * If you just want to \"defer\" task, that is to perform it right after currently\n * executing synchronous code ends (commonly achieved by `setTimeout(deferredTask, 0)`),\n * better choice will be the {@link asapScheduler} scheduler.\n *\n * ## Examples\n * Use async scheduler to delay task\n * ```ts\n * import { asyncScheduler } from 'rxjs';\n *\n * const task = () => console.log('it works!');\n *\n * asyncScheduler.schedule(task, 2000);\n *\n * // After 2 seconds logs:\n * // \"it works!\"\n * ```\n *\n * Use async scheduler to repeat task in intervals\n * ```ts\n * import { asyncScheduler } from 'rxjs';\n *\n * function task(state) {\n * console.log(state);\n * this.schedule(state + 1, 1000); // `this` references currently executing Action,\n * // which we reschedule with new state and delay\n * }\n *\n * asyncScheduler.schedule(task, 3000, 0);\n *\n * // Logs:\n * // 0 after 3s\n * // 1 after 4s\n * // 2 after 5s\n * // 3 after 6s\n * ```\n */\n\nexport const asyncScheduler = new AsyncScheduler(AsyncAction);\n\n/**\n * @deprecated Renamed to {@link asyncScheduler}. Will be removed in v8.\n */\nexport const async = asyncScheduler;\n", "import { AsyncAction } from './AsyncAction';\nimport { Subscription } from '../Subscription';\nimport { QueueScheduler } from './QueueScheduler';\nimport { SchedulerAction } from '../types';\nimport { TimerHandle } from './timerHandle';\n\nexport class QueueAction extends AsyncAction {\n constructor(protected scheduler: QueueScheduler, protected work: (this: SchedulerAction, state?: T) => void) {\n super(scheduler, work);\n }\n\n public schedule(state?: T, delay: number = 0): Subscription {\n if (delay > 0) {\n return super.schedule(state, delay);\n }\n this.delay = delay;\n this.state = state;\n this.scheduler.flush(this);\n return this;\n }\n\n public execute(state: T, delay: number): any {\n return delay > 0 || this.closed ? super.execute(state, delay) : this._execute(state, delay);\n }\n\n protected requestAsyncId(scheduler: QueueScheduler, id?: TimerHandle, delay: number = 0): TimerHandle {\n // If delay exists and is greater than 0, or if the delay is null (the\n // action wasn't rescheduled) but was originally scheduled as an async\n // action, then recycle as an async action.\n\n if ((delay != null && delay > 0) || (delay == null && this.delay > 0)) {\n return super.requestAsyncId(scheduler, id, delay);\n }\n\n // Otherwise flush the scheduler starting with this action.\n scheduler.flush(this);\n\n // HACK: In the past, this was returning `void`. However, `void` isn't a valid\n // `TimerHandle`, and generally the return value here isn't really used. So the\n // compromise is to return `0` which is both \"falsy\" and a valid `TimerHandle`,\n // as opposed to refactoring every other instanceo of `requestAsyncId`.\n return 0;\n }\n}\n", "import { AsyncScheduler } from './AsyncScheduler';\n\nexport class QueueScheduler extends AsyncScheduler {\n}\n", "import { QueueAction } from './QueueAction';\nimport { QueueScheduler } from './QueueScheduler';\n\n/**\n *\n * Queue Scheduler\n *\n * Put every next task on a queue, instead of executing it immediately\n *\n * `queue` scheduler, when used with delay, behaves the same as {@link asyncScheduler} scheduler.\n *\n * When used without delay, it schedules given task synchronously - executes it right when\n * it is scheduled. However when called recursively, that is when inside the scheduled task,\n * another task is scheduled with queue scheduler, instead of executing immediately as well,\n * that task will be put on a queue and wait for current one to finish.\n *\n * This means that when you execute task with `queue` scheduler, you are sure it will end\n * before any other task scheduled with that scheduler will start.\n *\n * ## Examples\n * Schedule recursively first, then do something\n * ```ts\n * import { queueScheduler } from 'rxjs';\n *\n * queueScheduler.schedule(() => {\n * queueScheduler.schedule(() => console.log('second')); // will not happen now, but will be put on a queue\n *\n * console.log('first');\n * });\n *\n * // Logs:\n * // \"first\"\n * // \"second\"\n * ```\n *\n * Reschedule itself recursively\n * ```ts\n * import { queueScheduler } from 'rxjs';\n *\n * queueScheduler.schedule(function(state) {\n * if (state !== 0) {\n * console.log('before', state);\n * this.schedule(state - 1); // `this` references currently executing Action,\n * // which we reschedule with new state\n * console.log('after', state);\n * }\n * }, 0, 3);\n *\n * // In scheduler that runs recursively, you would expect:\n * // \"before\", 3\n * // \"before\", 2\n * // \"before\", 1\n * // \"after\", 1\n * // \"after\", 2\n * // \"after\", 3\n *\n * // But with queue it logs:\n * // \"before\", 3\n * // \"after\", 3\n * // \"before\", 2\n * // \"after\", 2\n * // \"before\", 1\n * // \"after\", 1\n * ```\n */\n\nexport const queueScheduler = new QueueScheduler(QueueAction);\n\n/**\n * @deprecated Renamed to {@link queueScheduler}. Will be removed in v8.\n */\nexport const queue = queueScheduler;\n", "import { AsyncAction } from './AsyncAction';\nimport { AnimationFrameScheduler } from './AnimationFrameScheduler';\nimport { SchedulerAction } from '../types';\nimport { animationFrameProvider } from './animationFrameProvider';\nimport { TimerHandle } from './timerHandle';\n\nexport class AnimationFrameAction extends AsyncAction {\n constructor(protected scheduler: AnimationFrameScheduler, protected work: (this: SchedulerAction, state?: T) => void) {\n super(scheduler, work);\n }\n\n protected requestAsyncId(scheduler: AnimationFrameScheduler, id?: TimerHandle, delay: number = 0): TimerHandle {\n // If delay is greater than 0, request as an async action.\n if (delay !== null && delay > 0) {\n return super.requestAsyncId(scheduler, id, delay);\n }\n // Push the action to the end of the scheduler queue.\n scheduler.actions.push(this);\n // If an animation frame has already been requested, don't request another\n // one. If an animation frame hasn't been requested yet, request one. Return\n // the current animation frame request id.\n return scheduler._scheduled || (scheduler._scheduled = animationFrameProvider.requestAnimationFrame(() => scheduler.flush(undefined)));\n }\n\n protected recycleAsyncId(scheduler: AnimationFrameScheduler, id?: TimerHandle, delay: number = 0): TimerHandle | undefined {\n // If delay exists and is greater than 0, or if the delay is null (the\n // action wasn't rescheduled) but was originally scheduled as an async\n // action, then recycle as an async action.\n if (delay != null ? delay > 0 : this.delay > 0) {\n return super.recycleAsyncId(scheduler, id, delay);\n }\n // If the scheduler queue has no remaining actions with the same async id,\n // cancel the requested animation frame and set the scheduled flag to\n // undefined so the next AnimationFrameAction will request its own.\n const { actions } = scheduler;\n if (id != null && actions[actions.length - 1]?.id !== id) {\n animationFrameProvider.cancelAnimationFrame(id as number);\n scheduler._scheduled = undefined;\n }\n // Return undefined so the action knows to request a new async id if it's rescheduled.\n return undefined;\n }\n}\n", "import { AsyncAction } from './AsyncAction';\nimport { AsyncScheduler } from './AsyncScheduler';\n\nexport class AnimationFrameScheduler extends AsyncScheduler {\n public flush(action?: AsyncAction): void {\n this._active = true;\n // The async id that effects a call to flush is stored in _scheduled.\n // Before executing an action, it's necessary to check the action's async\n // id to determine whether it's supposed to be executed in the current\n // flush.\n // Previous implementations of this method used a count to determine this,\n // but that was unsound, as actions that are unsubscribed - i.e. cancelled -\n // are removed from the actions array and that can shift actions that are\n // scheduled to be executed in a subsequent flush into positions at which\n // they are executed within the current flush.\n const flushId = this._scheduled;\n this._scheduled = undefined;\n\n const { actions } = this;\n let error: any;\n action = action || actions.shift()!;\n\n do {\n if ((error = action.execute(action.state, action.delay))) {\n break;\n }\n } while ((action = actions[0]) && action.id === flushId && actions.shift());\n\n this._active = false;\n\n if (error) {\n while ((action = actions[0]) && action.id === flushId && actions.shift()) {\n action.unsubscribe();\n }\n throw error;\n }\n }\n}\n", "import { AnimationFrameAction } from './AnimationFrameAction';\nimport { AnimationFrameScheduler } from './AnimationFrameScheduler';\n\n/**\n *\n * Animation Frame Scheduler\n *\n * Perform task when `window.requestAnimationFrame` would fire\n *\n * When `animationFrame` scheduler is used with delay, it will fall back to {@link asyncScheduler} scheduler\n * behaviour.\n *\n * Without delay, `animationFrame` scheduler can be used to create smooth browser animations.\n * It makes sure scheduled task will happen just before next browser content repaint,\n * thus performing animations as efficiently as possible.\n *\n * ## Example\n * Schedule div height animation\n * ```ts\n * // html: