{
  "nbformat": 4,
  "nbformat_minor": 0,
  "metadata": {
    "colab": {
      "provenance": [],
      "gpuType": "T4"
    },
    "kernelspec": {
      "name": "python3",
      "display_name": "Python 3"
    },
    "language_info": {
      "name": "python"
    },
    "accelerator": "GPU",
    "widgets": {
      "application/vnd.jupyter.widget-state+json": {
        "105fb75dea9c468d98e36ba48b91d379": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HBoxModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HBoxModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HBoxView",
            "box_style": "",
            "children": [
              "IPY_MODEL_0dbda088ec90412e959e6290392d97dc",
              "IPY_MODEL_46f8cb9a4775444db4a2a7b2e0a72bf0",
              "IPY_MODEL_851b3bbe139244aabede18801d4a923a"
            ],
            "layout": "IPY_MODEL_21039746a8844d88883a517406891d50"
          }
        },
        "0dbda088ec90412e959e6290392d97dc": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_d3e0703dd70e486fa923ac2c37b1c879",
            "placeholder": "​",
            "style": "IPY_MODEL_4f7ad333a1af45f48e174d01b34dc561",
            "value": "tokenizer_config.json: 100%"
          }
        },
        "46f8cb9a4775444db4a2a7b2e0a72bf0": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "FloatProgressModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "FloatProgressModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "ProgressView",
            "bar_style": "success",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_ac84da5392dd4778b3779d85c182f357",
            "max": 1289,
            "min": 0,
            "orientation": "horizontal",
            "style": "IPY_MODEL_1388398ab4f74ab0831100350a1d99a7",
            "value": 1289
          }
        },
        "851b3bbe139244aabede18801d4a923a": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_37a7646419974760845d692d0ec9ee44",
            "placeholder": "​",
            "style": "IPY_MODEL_8dca6c4e6aee4c708b3760ab5d103c28",
            "value": " 1.29k/1.29k [00:00&lt;00:00, 33.1kB/s]"
          }
        },
        "21039746a8844d88883a517406891d50": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "d3e0703dd70e486fa923ac2c37b1c879": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "4f7ad333a1af45f48e174d01b34dc561": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "ac84da5392dd4778b3779d85c182f357": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "1388398ab4f74ab0831100350a1d99a7": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "ProgressStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "ProgressStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "bar_color": null,
            "description_width": ""
          }
        },
        "37a7646419974760845d692d0ec9ee44": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "8dca6c4e6aee4c708b3760ab5d103c28": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "b1678de1646e494b863cf49efa51e4b4": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HBoxModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HBoxModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HBoxView",
            "box_style": "",
            "children": [
              "IPY_MODEL_251e1602b2a740ee8f7f8889ee2149a6",
              "IPY_MODEL_c2f46f41d87b4a878dbef56baefebe84",
              "IPY_MODEL_fc3c9dac87ca4c858654c1b1a50890ac"
            ],
            "layout": "IPY_MODEL_d6514f03f917423fbf93a551102c28fe"
          }
        },
        "251e1602b2a740ee8f7f8889ee2149a6": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_19c073826a644890989d6c453e561b8d",
            "placeholder": "​",
            "style": "IPY_MODEL_63e88dbb26044265b6a75bb7112e84c7",
            "value": "tokenizer.model: 100%"
          }
        },
        "c2f46f41d87b4a878dbef56baefebe84": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "FloatProgressModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "FloatProgressModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "ProgressView",
            "bar_style": "success",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_fd5c322c28e64b99a8c0e877a79e59c2",
            "max": 499723,
            "min": 0,
            "orientation": "horizontal",
            "style": "IPY_MODEL_8f737ff411d040dc8bf4aeb95e8c3cd6",
            "value": 499723
          }
        },
        "fc3c9dac87ca4c858654c1b1a50890ac": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_1c5b20fa7e86463db5b40a3226d28a57",
            "placeholder": "​",
            "style": "IPY_MODEL_95568254f93e4bd29c151647bcbc36e9",
            "value": " 500k/500k [00:00&lt;00:00, 7.28MB/s]"
          }
        },
        "d6514f03f917423fbf93a551102c28fe": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "19c073826a644890989d6c453e561b8d": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "63e88dbb26044265b6a75bb7112e84c7": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "fd5c322c28e64b99a8c0e877a79e59c2": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "8f737ff411d040dc8bf4aeb95e8c3cd6": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "ProgressStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "ProgressStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "bar_color": null,
            "description_width": ""
          }
        },
        "1c5b20fa7e86463db5b40a3226d28a57": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "95568254f93e4bd29c151647bcbc36e9": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "bfeb2fe64096492cb11f6e6548cd4085": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HBoxModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HBoxModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HBoxView",
            "box_style": "",
            "children": [
              "IPY_MODEL_e327c7dfbb21432fb8637ca693623878",
              "IPY_MODEL_c29e202a0d57417dab74485f36bf4259",
              "IPY_MODEL_7a05e30b75fd499998fe2d2e1ba49de3"
            ],
            "layout": "IPY_MODEL_17c3b47d4ba24f6fbddf947123b3bfbf"
          }
        },
        "e327c7dfbb21432fb8637ca693623878": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_909d46c781d64f0f85c6eb6b08a0505c",
            "placeholder": "​",
            "style": "IPY_MODEL_f335d3386bf74e928678175451e58904",
            "value": "tokenizer.json: 100%"
          }
        },
        "c29e202a0d57417dab74485f36bf4259": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "FloatProgressModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "FloatProgressModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "ProgressView",
            "bar_style": "success",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_000e1e07eaf44009a643eebb1bbd4333",
            "max": 1842767,
            "min": 0,
            "orientation": "horizontal",
            "style": "IPY_MODEL_d5c3d3d16d174bf39dbffb2320b7143a",
            "value": 1842767
          }
        },
        "7a05e30b75fd499998fe2d2e1ba49de3": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_78273c7bfe144d439a2a16d30652c6d5",
            "placeholder": "​",
            "style": "IPY_MODEL_a8649eb6dcbf48d1812b4b6076a5edd3",
            "value": " 1.84M/1.84M [00:00&lt;00:00, 2.62MB/s]"
          }
        },
        "17c3b47d4ba24f6fbddf947123b3bfbf": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "909d46c781d64f0f85c6eb6b08a0505c": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "f335d3386bf74e928678175451e58904": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "000e1e07eaf44009a643eebb1bbd4333": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "d5c3d3d16d174bf39dbffb2320b7143a": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "ProgressStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "ProgressStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "bar_color": null,
            "description_width": ""
          }
        },
        "78273c7bfe144d439a2a16d30652c6d5": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "a8649eb6dcbf48d1812b4b6076a5edd3": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "54c3b3b1aaf540568618094a4dd7c2de": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HBoxModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HBoxModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HBoxView",
            "box_style": "",
            "children": [
              "IPY_MODEL_fc14cbf3ab3e412a9842f818596ea739",
              "IPY_MODEL_6374f0370842469b95f9d316416fd71e",
              "IPY_MODEL_40c8035883f44aecb7e2373861f7752e"
            ],
            "layout": "IPY_MODEL_9af36ae1fa954607a5fd9207d52bae10"
          }
        },
        "fc14cbf3ab3e412a9842f818596ea739": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_55d85509469541e09fdb67c16196c72a",
            "placeholder": "​",
            "style": "IPY_MODEL_f4da21a5eeef4b4bb278d9f03ee6061b",
            "value": "special_tokens_map.json: 100%"
          }
        },
        "6374f0370842469b95f9d316416fd71e": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "FloatProgressModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "FloatProgressModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "ProgressView",
            "bar_style": "success",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_cc4dc428f3734b7685e941fac87ec5ba",
            "max": 551,
            "min": 0,
            "orientation": "horizontal",
            "style": "IPY_MODEL_57844d99675a49658b8e002b61e2fc12",
            "value": 551
          }
        },
        "40c8035883f44aecb7e2373861f7752e": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_794b056f88214be3a9c662785d26d009",
            "placeholder": "​",
            "style": "IPY_MODEL_c124d1ce85d54577be6572e2b926354f",
            "value": " 551/551 [00:00&lt;00:00, 7.81kB/s]"
          }
        },
        "9af36ae1fa954607a5fd9207d52bae10": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "55d85509469541e09fdb67c16196c72a": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "f4da21a5eeef4b4bb278d9f03ee6061b": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "cc4dc428f3734b7685e941fac87ec5ba": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "57844d99675a49658b8e002b61e2fc12": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "ProgressStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "ProgressStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "bar_color": null,
            "description_width": ""
          }
        },
        "794b056f88214be3a9c662785d26d009": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "c124d1ce85d54577be6572e2b926354f": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "b4c10becbd7845a2828ef22975db1e74": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HBoxModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HBoxModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HBoxView",
            "box_style": "",
            "children": [
              "IPY_MODEL_513868ac37684323b30a05bb236f3c40",
              "IPY_MODEL_074bc3c411ab4be9b3101780efb62d33",
              "IPY_MODEL_e085f52e520d431692e03b3d6b9890ed"
            ],
            "layout": "IPY_MODEL_27578130937a49379b157b92c52aee9b"
          }
        },
        "513868ac37684323b30a05bb236f3c40": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_aaaa468f9c174c948a6a74204af94440",
            "placeholder": "​",
            "style": "IPY_MODEL_ebdc5604a4394bfa9bd847b583f02940",
            "value": "config.json: 100%"
          }
        },
        "074bc3c411ab4be9b3101780efb62d33": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "FloatProgressModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "FloatProgressModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "ProgressView",
            "bar_style": "success",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_50e0afcbf70b4edea8d6010587ec3bcc",
            "max": 608,
            "min": 0,
            "orientation": "horizontal",
            "style": "IPY_MODEL_921ba7da88b742c19625c1aa1a32a99e",
            "value": 608
          }
        },
        "e085f52e520d431692e03b3d6b9890ed": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_7a38f868e970400b9a4a8e595cb6f08f",
            "placeholder": "​",
            "style": "IPY_MODEL_a4e7ed98efe844a5a447faca24bbf326",
            "value": " 608/608 [00:00&lt;00:00, 67.3kB/s]"
          }
        },
        "27578130937a49379b157b92c52aee9b": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "aaaa468f9c174c948a6a74204af94440": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "ebdc5604a4394bfa9bd847b583f02940": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "50e0afcbf70b4edea8d6010587ec3bcc": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "921ba7da88b742c19625c1aa1a32a99e": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "ProgressStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "ProgressStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "bar_color": null,
            "description_width": ""
          }
        },
        "7a38f868e970400b9a4a8e595cb6f08f": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "a4e7ed98efe844a5a447faca24bbf326": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "339f3ea0ab1b4594bd38d5e274123c1e": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HBoxModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HBoxModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HBoxView",
            "box_style": "",
            "children": [
              "IPY_MODEL_d2c6320b1edd41829ada662c18dc5a93",
              "IPY_MODEL_ce14d381e1d3436caa3a911f1f70a9d9",
              "IPY_MODEL_fe880a41914943398ee851e1932c5c87"
            ],
            "layout": "IPY_MODEL_931943c3c69a45969d829d89e437a937"
          }
        },
        "d2c6320b1edd41829ada662c18dc5a93": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_5a579c5bb74243f4b4ff0daaa3fdee8a",
            "placeholder": "​",
            "style": "IPY_MODEL_3ac155c48bac45bba348ac0853582c24",
            "value": "model.safetensors: 100%"
          }
        },
        "ce14d381e1d3436caa3a911f1f70a9d9": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "FloatProgressModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "FloatProgressModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "ProgressView",
            "bar_style": "success",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_037d002dff9648cc921edd18e024bf5e",
            "max": 2200119864,
            "min": 0,
            "orientation": "horizontal",
            "style": "IPY_MODEL_c6e4a72224e0486f81763399326c0e1a",
            "value": 2200119864
          }
        },
        "fe880a41914943398ee851e1932c5c87": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_89a9dfadcda34370a8b61ac2374514a8",
            "placeholder": "​",
            "style": "IPY_MODEL_7dc638cdd4d64236aea2edd3ef286627",
            "value": " 2.20G/2.20G [00:18&lt;00:00, 245MB/s]"
          }
        },
        "931943c3c69a45969d829d89e437a937": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "5a579c5bb74243f4b4ff0daaa3fdee8a": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "3ac155c48bac45bba348ac0853582c24": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "037d002dff9648cc921edd18e024bf5e": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "c6e4a72224e0486f81763399326c0e1a": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "ProgressStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "ProgressStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "bar_color": null,
            "description_width": ""
          }
        },
        "89a9dfadcda34370a8b61ac2374514a8": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "7dc638cdd4d64236aea2edd3ef286627": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "7df660e81b844248a6deaed0decaef23": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HBoxModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HBoxModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HBoxView",
            "box_style": "",
            "children": [
              "IPY_MODEL_c93a8faf318049f68729f2e34adef2b2",
              "IPY_MODEL_2fe5bf5c529249daaea5517dc00a9c8f",
              "IPY_MODEL_953fb98fe8c84f6aabba26898a98e664"
            ],
            "layout": "IPY_MODEL_172d023d55814b02a96f383ab42ca5e1"
          }
        },
        "c93a8faf318049f68729f2e34adef2b2": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_21c634a44479472fbeb4d20af25a35c6",
            "placeholder": "​",
            "style": "IPY_MODEL_79a9c02dd5bf43cc92019014e1a256fd",
            "value": "generation_config.json: 100%"
          }
        },
        "2fe5bf5c529249daaea5517dc00a9c8f": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "FloatProgressModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "FloatProgressModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "ProgressView",
            "bar_style": "success",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_49062ef858464c9490cb50c509dc99de",
            "max": 124,
            "min": 0,
            "orientation": "horizontal",
            "style": "IPY_MODEL_5f2ee295c18142a2b398afac5cd4608b",
            "value": 124
          }
        },
        "953fb98fe8c84f6aabba26898a98e664": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "HTMLModel",
          "model_module_version": "1.5.0",
          "state": {
            "_dom_classes": [],
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "HTMLModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/controls",
            "_view_module_version": "1.5.0",
            "_view_name": "HTMLView",
            "description": "",
            "description_tooltip": null,
            "layout": "IPY_MODEL_40c36240b27d4cc889f312a2129021aa",
            "placeholder": "​",
            "style": "IPY_MODEL_d4d9f8b3fcbb4490abf85183f91b2601",
            "value": " 124/124 [00:00&lt;00:00, 13.6kB/s]"
          }
        },
        "172d023d55814b02a96f383ab42ca5e1": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "21c634a44479472fbeb4d20af25a35c6": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "79a9c02dd5bf43cc92019014e1a256fd": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        },
        "49062ef858464c9490cb50c509dc99de": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "5f2ee295c18142a2b398afac5cd4608b": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "ProgressStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "ProgressStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "bar_color": null,
            "description_width": ""
          }
        },
        "40c36240b27d4cc889f312a2129021aa": {
          "model_module": "@jupyter-widgets/base",
          "model_name": "LayoutModel",
          "model_module_version": "1.2.0",
          "state": {
            "_model_module": "@jupyter-widgets/base",
            "_model_module_version": "1.2.0",
            "_model_name": "LayoutModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "LayoutView",
            "align_content": null,
            "align_items": null,
            "align_self": null,
            "border": null,
            "bottom": null,
            "display": null,
            "flex": null,
            "flex_flow": null,
            "grid_area": null,
            "grid_auto_columns": null,
            "grid_auto_flow": null,
            "grid_auto_rows": null,
            "grid_column": null,
            "grid_gap": null,
            "grid_row": null,
            "grid_template_areas": null,
            "grid_template_columns": null,
            "grid_template_rows": null,
            "height": null,
            "justify_content": null,
            "justify_items": null,
            "left": null,
            "margin": null,
            "max_height": null,
            "max_width": null,
            "min_height": null,
            "min_width": null,
            "object_fit": null,
            "object_position": null,
            "order": null,
            "overflow": null,
            "overflow_x": null,
            "overflow_y": null,
            "padding": null,
            "right": null,
            "top": null,
            "visibility": null,
            "width": null
          }
        },
        "d4d9f8b3fcbb4490abf85183f91b2601": {
          "model_module": "@jupyter-widgets/controls",
          "model_name": "DescriptionStyleModel",
          "model_module_version": "1.5.0",
          "state": {
            "_model_module": "@jupyter-widgets/controls",
            "_model_module_version": "1.5.0",
            "_model_name": "DescriptionStyleModel",
            "_view_count": null,
            "_view_module": "@jupyter-widgets/base",
            "_view_module_version": "1.2.0",
            "_view_name": "StyleView",
            "description_width": ""
          }
        }
      }
    }
  },
  "cells": [
    {
      "cell_type": "markdown",
      "source": [
        "This code is meant to recreate Fig-2 for submission:\n",
        "\n",
        "---\n",
        "\n",
        "\"**How Much Context Does Natural Language Actually Require?\n",
        "An Analysis Using LLMs as Statistical Oracles**\"\n",
        "\n",
        "---\n",
        "which evaluates the Minimal Context Length (**MCL**) required for context/next-token pairs extracted from stories sampled from TinyStories. We use TinyLlama as the reference oracle and run the experiments in a minimal configuration to enable faster runtimes on free Colab instances. Note that actual runtime may vary depending on the number of sampled stories. I used ChatGPT to help draft documentation and improve code readability. For questions about this code or alternative runtime setups used for the other figures in the paper, feel free to contact me by email!\n"
      ],
      "metadata": {
        "id": "cxhyUtYIyxip"
      }
    },
    {
      "cell_type": "code",
      "source": [
        "# Initial imports\n",
        "import torch\n",
        "import os\n",
        "from transformers import AutoTokenizer, AutoModelForCausalLM\n",
        "from datasets import load_dataset, disable_caching\n",
        "from tqdm import tqdm\n",
        "import matplotlib.pyplot as plt\n",
        "import json\n",
        "import numpy as np\n",
        "from contextlib import nullcontext\n",
        "import torch.nn.functional as F\n",
        "import shutil\n",
        "import tempfile\n",
        "import requests\n",
        "import numpy as np\n",
        "from scipy.optimize import curve_fit\n",
        "import matplotlib.pyplot as plt\n",
        "\n",
        "# Disable all caching for datasets\n",
        "disable_caching()\n"
      ],
      "metadata": {
        "id": "tlsrybrQvPQv"
      },
      "execution_count": 1,
      "outputs": []
    },
    {
      "cell_type": "code",
      "source": [
        "\n",
        "def analyze_prediction_confidence_fast(model, input_ids, conf_thresh=0.2):\n",
        "    \"\"\"\n",
        "    Quickly identify positions in a token sequence where the model confidently predicts the correct next token.\n",
        "\n",
        "    This function runs a single forward pass over the input, compares the top-1 prediction with the ground truth,\n",
        "    and considers it confident if the probability margin between top-1 and top-2 predictions exceeds `conf_thresh`.\n",
        "    Meant to filter out non-confident tokens for the MCL analysis.\n",
        "\n",
        "    Args:\n",
        "        model (PreTrainedModel): A causal language model.\n",
        "        input_ids (torch.Tensor): 1D tensor of token IDs (not batched).\n",
        "        conf_thresh (float): Minimum required margin (p1 - p2) between top-1 and top-2 token probabilities\n",
        "                             to be considered confident.\n",
        "\n",
        "    Returns:\n",
        "        tuple:\n",
        "            - correct_conf (List[int]): Indices (1-based) of tokens where the model predicted the correct next token\n",
        "                                        with high confidence.\n",
        "            - input_ids[0] (torch.Tensor): The original token sequence (on CUDA).\n",
        "    \"\"\"\n",
        "    input_ids = input_ids.unsqueeze(0).cuda()\n",
        "    with torch.no_grad():\n",
        "        logits = model(input_ids=input_ids, use_cache=False).logits[:, :-1, :]\n",
        "        probs = torch.nn.functional.softmax(logits, dim=-1).squeeze(0).cpu()\n",
        "    targets = input_ids[0, 1:].cpu()\n",
        "    top_probs, top_ids = torch.topk(probs, 2, dim=-1)\n",
        "    correct_conf = []\n",
        "    # Iterate through the tokens in the dataset\n",
        "    for i in range(len(targets)):\n",
        "        gold = targets[i].item()\n",
        "        top1, top2 = top_ids[i]\n",
        "        p1, p2 = top_probs[i]\n",
        "        # This is where we check if the context is correctly and confidently predicted by the model\n",
        "        if top1.item() == gold and (p1.item() - p2.item()) > conf_thresh:\n",
        "            correct_conf.append(i + 1)\n",
        "    return correct_conf, input_ids[0]\n",
        "\n",
        "\n",
        "def stratified_sample(indices, seq_len, num_samples=100, bins=10):\n",
        "    \"\"\"\n",
        "    Perform stratified sampling over a sequence by dividing it into bins and sampling uniformly from each bin.\n",
        "    Given that we do a filtering using 'analyze_prediction_confidence_fast', we want to make sure we still sample equallt\n",
        "    across the document.\n",
        "\n",
        "    This function ensures that sampled indices are distributed across the sequence length by:\n",
        "    - Dividing the sequence into `bins` equal-length intervals,\n",
        "    - Sampling approximately `num_samples // bins` from each bin,\n",
        "    - Filling any remaining quota by sampling from unused indices if needed.\n",
        "\n",
        "    Args:\n",
        "        indices (array-like): The candidate indices to sample from (e.g., token positions).\n",
        "        seq_len (int): Total sequence length (used to define bin edges).\n",
        "        num_samples (int): Total number of samples to select.\n",
        "        bins (int): Number of bins to stratify across.\n",
        "\n",
        "    Returns:\n",
        "        List[int]: Sorted list of sampled indices stratified across the sequence.\n",
        "    \"\"\"\n",
        "    indices = np.array(indices)\n",
        "    sampled_indices = []\n",
        "    bin_edges = np.linspace(0, seq_len, bins + 1, dtype=int)\n",
        "    samples_per_bin = num_samples // bins\n",
        "    for i in range(bins):\n",
        "        start, end = bin_edges[i], bin_edges[i + 1]\n",
        "        bin_members = indices[(indices >= start) & (indices < end)]\n",
        "        if len(bin_members) > 0:\n",
        "            num = min(samples_per_bin, len(bin_members))\n",
        "            sampled = np.random.choice(bin_members, size=num, replace=False)\n",
        "            sampled_indices.extend(sampled.tolist())\n",
        "    if len(sampled_indices) < num_samples:\n",
        "        extra_needed = num_samples - len(sampled_indices)\n",
        "        extra_pool = np.setdiff1d(indices, sampled_indices)\n",
        "        if len(extra_pool) >= extra_needed:\n",
        "            extra = np.random.choice(extra_pool, size=extra_needed, replace=False)\n",
        "            sampled_indices.extend(extra.tolist())\n",
        "    return sorted(sampled_indices)\n"
      ],
      "metadata": {
        "id": "oY_G-GhcvVyr"
      },
      "execution_count": 2,
      "outputs": []
    },
    {
      "cell_type": "code",
      "source": [
        "\n",
        "def find_tokenwise_true_context_lengths(model, tokenizer, input_ids, target_indices, threshold=0.2,\n",
        "                                        step_size=8, min_len=32, use_masking=True,\n",
        "                                        batch_sizes={32: 1, 512: 1}):\n",
        "    \"\"\"\n",
        "    Determine the minimum context length required to confidently predict each token in a sequence.\n",
        "\n",
        "    For each target token index in `target_indices`, this function gradually increases the length of context\n",
        "    available to the model (starting from `min_len` up to `idx`), and checks whether the model's top-1 predicted\n",
        "    token matches the ground truth token with a confidence margin greater than `threshold`.\n",
        "\n",
        "    Args:\n",
        "        model (PreTrainedModel): A causal language model.\n",
        "        tokenizer (PreTrainedTokenizer): The tokenizer corresponding to the model.\n",
        "        input_ids (torch.Tensor): 1D tensor of tokenized input.\n",
        "        target_indices (list[int]): List of token indices for which to determine context sensitivity.\n",
        "        threshold (float): Minimum margin (difference between top-1 and top-2 token probs) to be considered confident.\n",
        "        step_size (int): Increment size for context window during probing.\n",
        "        min_len (int): Minimum context length to begin checking from.\n",
        "        use_masking (bool): If True, uses attention masks to simulate different context sizes.\n",
        "                            If False, pads context to simulate variable-length inputs.\n",
        "        batch_sizes (dict): Mapping from context length to batch size, to control memory usage.\n",
        "\n",
        "    Returns:\n",
        "        dict: Mapping from token index to a dict with key 'conf_ctx_len' indicating\n",
        "              the minimum context length where the token is confidently predicted.\n",
        "              If no such context length is found, value is None.\n",
        "    \"\"\"\n",
        "\n",
        "    results = {}\n",
        "    input_ids = input_ids.cuda()\n",
        "    for idx in tqdm(target_indices, desc=\"Checking tokens for context lengths\"):\n",
        "\n",
        "        max_ctx = idx\n",
        "        if max_ctx < min_len:\n",
        "            continue\n",
        "        # Split the context into appropriate subcontext, incrementing by step_size\n",
        "        candidate_lens = list(range(min_len, max_ctx + 1, step_size))\n",
        "        actual_token = input_ids[idx].item()\n",
        "        conf_ctx_len = None\n",
        "\n",
        "        def get_batch_size(l):\n",
        "            \"\"\"\n",
        "            Helper fucntion used to control memory usage when batching on smaller memory GPUs.\n",
        "            \"\"\"\n",
        "            for k in sorted(batch_sizes.keys(), reverse=True):\n",
        "                if l >= k:\n",
        "                    return batch_sizes[k]\n",
        "            return batch_sizes[min(batch_sizes)]\n",
        "\n",
        "        with torch.no_grad():\n",
        "\n",
        "            # Go through different subcontext lengths for one context\n",
        "            for batch_start in range(0, len(candidate_lens), get_batch_size(candidate_lens[-1])):\n",
        "                batch_lens = candidate_lens[batch_start: batch_start + get_batch_size(candidate_lens[-1])]\n",
        "\n",
        "                # If we prefer to use the correct positional encoding, use attention masking to create the subcontexts\n",
        "                if use_masking:\n",
        "                    context = input_ids[:idx].unsqueeze(0).repeat(len(batch_lens), 1)\n",
        "                    attention_mask = torch.zeros_like(context)\n",
        "                    for i, ctx_len in enumerate(batch_lens):\n",
        "                        attention_mask[i, -ctx_len:] = 1\n",
        "                    context = context.cuda()\n",
        "                    attention_mask = attention_mask.cuda()\n",
        "                # Or we could just pass in the shortened context itself\n",
        "                else:\n",
        "                    max_len = max(batch_lens)\n",
        "                    padded_contexts = []\n",
        "                    for l in batch_lens:\n",
        "                        slice = input_ids[idx - l:idx]\n",
        "                        pad_len = max_len - slice.size(0)\n",
        "                        padded = F.pad(slice, (pad_len, 0), value=tokenizer.pad_token_id)\n",
        "                        padded_contexts.append(padded)\n",
        "                    context = torch.stack(padded_contexts)\n",
        "                    attention_mask = None\n",
        "\n",
        "                outputs = model(input_ids=context, attention_mask=attention_mask)\n",
        "                logits = outputs.logits[:, -1, :]\n",
        "                probs = torch.nn.functional.softmax(logits, dim=-1).cpu()\n",
        "\n",
        "                # Go through the subcontext logits in the batch to evaluate MCL\n",
        "                for i, ctx_len in enumerate(batch_lens):\n",
        "                    top_probs, top_ids = torch.topk(probs[i], 2)\n",
        "                    top1, top2 = top_ids.tolist()\n",
        "                    p1, p2 = top_probs.tolist()\n",
        "                    margin = p1 - p2\n",
        "                    # check if a subcontext matches the requirements of MCL\n",
        "                    if conf_ctx_len is None and top1 == actual_token and margin > threshold:\n",
        "                        conf_ctx_len = ctx_len\n",
        "\n",
        "                del outputs, logits, probs\n",
        "                torch.cuda.empty_cache()\n",
        "\n",
        "                # To save computation, stop when the first correct subcontext is selected\n",
        "                if conf_ctx_len is not None:\n",
        "                    break\n",
        "\n",
        "        results[idx] = {\"conf_ctx_len\": conf_ctx_len}\n",
        "    return results\n",
        "\n"
      ],
      "metadata": {
        "id": "Ragu4YiGvmak"
      },
      "execution_count": 3,
      "outputs": []
    },
    {
      "cell_type": "code",
      "source": [
        "def plot_confidence_pie(total_confident_tokens, total_tokens, processed_stories, save_path):\n",
        "    \"\"\"\n",
        "    Plot a pie chart showing the proportion of confidently predicted tokens.\n",
        "\n",
        "    Args:\n",
        "        total_confident_tokens (int): Number of tokens confidently predicted.\n",
        "        total_tokens (int): Total number of predicted tokens.\n",
        "        processed_stories (int): Number of processed stories.\n",
        "        save_path (str): Output path for the plot.\n",
        "    \"\"\"\n",
        "    confident_pct = total_confident_tokens / total_tokens if total_tokens > 0 else 0\n",
        "\n",
        "    plt.figure(figsize=(6, 6))\n",
        "    plt.pie([confident_pct, 1 - confident_pct],\n",
        "            labels=[f\"Confident ({confident_pct:.1%})\", \"Other\"],\n",
        "            colors=[\"green\", \"lightgray\"], autopct=\"%1.1f%%\")\n",
        "    plt.title(f\"Token Classification After {processed_stories} Stories\")\n",
        "    plt.tight_layout()\n",
        "    plt.savefig(save_path)\n",
        "    plt.close()\n",
        "\n",
        "\n",
        "\n",
        "def plot_context_length_histogram(conf_lens, save_path):\n",
        "    \"\"\"\n",
        "    Plot histogram of minimum context lengths and overlay a power-law fit.\n",
        "\n",
        "    Args:\n",
        "        conf_lens (List[int]): List of confident context lengths.\n",
        "        save_path (str): Output path for the plot.\n",
        "    \"\"\"\n",
        "    if not conf_lens:\n",
        "        print(\"⚠️ No confident context lengths to plot.\")\n",
        "        return\n",
        "\n",
        "    plt.figure(figsize=(8, 5))\n",
        "    bins = list(range(32, max(conf_lens) + 16, 16))\n",
        "    hist_vals, bin_edges = np.histogram(conf_lens, bins=bins)\n",
        "    plt.hist(conf_lens, bins=bins, color=\"green\", edgecolor=\"black\", alpha=0.8)\n",
        "    plt.yscale(\"log\")\n",
        "\n",
        "    # === Power law fit ===\n",
        "    def power_law(x, a, b):\n",
        "        return a * x ** (-b)\n",
        "\n",
        "    fit_x = 0.5 * (bin_edges[1:-1] + bin_edges[2:])\n",
        "    fit_y = hist_vals[1:]\n",
        "    valid = fit_y > 0\n",
        "    fit_x = fit_x[valid]\n",
        "    fit_y = fit_y[valid]\n",
        "\n",
        "    b_val = None\n",
        "    if len(fit_x) >= 2:\n",
        "        log_x = np.log(fit_x)\n",
        "        log_y = np.log(fit_y)\n",
        "        try:\n",
        "            popt, _ = curve_fit(lambda x, a, b: a + b * x, log_x, log_y)\n",
        "            a, b = popt\n",
        "            b_val = b\n",
        "            x_fit = np.linspace(min(fit_x), max(fit_x), 100)\n",
        "            y_fit = np.exp(a) * x_fit ** b\n",
        "            plt.plot(x_fit, y_fit, color='red', linewidth=2, label='Power-law fit')\n",
        "        except Exception as e:\n",
        "            print(f\"⚠️ Fit failed: {e}\")\n",
        "\n",
        "    # === Annotations ===\n",
        "    pct_below_64 = int(100 * sum(x <= 64 for x in conf_lens) / len(conf_lens) + 1)\n",
        "    plt.text(0.98, 0.95,\n",
        "             rf\"$p(l_i \\leq 64) = {pct_below_64:.1f}\\%$\",\n",
        "             transform=plt.gca().transAxes,\n",
        "             fontsize=13, ha='right', va='top')\n",
        "    if b_val is not None:\n",
        "        plt.text(0.98, 0.8,\n",
        "                 rf\"$\\hat{{b}} = {b_val:.2f}$\",\n",
        "                 transform=plt.gca().transAxes,\n",
        "                 fontsize=12, ha='right', va='top', color='red')\n",
        "\n",
        "    plt.xlabel(\"Minimum Context Length (MCL)\", fontsize=14)\n",
        "    plt.ylabel(\"Token Count (log scale)\", fontsize=14)\n",
        "    plt.title(\"Confident Tokens: Context Length Distribution\", fontsize=15)\n",
        "    plt.tight_layout()\n",
        "    plt.savefig(save_path)\n",
        "    plt.close()\n"
      ],
      "metadata": {
        "id": "DrfUvjMrxWjU"
      },
      "execution_count": 4,
      "outputs": []
    },
    {
      "cell_type": "code",
      "source": [
        "\n",
        "def run_context_analysis(model_name=\"tinyllama\", num_stories=1, min_tokens=1024,\n",
        "                         conf_thresh=0.2, num_samples_per_doc=100, output_dir=\"./output\"):\n",
        "    \"\"\"\n",
        "    Run context length sensitivity analysis on a TinyStories-style dataset using a causal LM.\n",
        "\n",
        "    For each story:\n",
        "    - Select confident token positions based on prediction margin.\n",
        "    - Stratify and sample tokens.\n",
        "    - Estimate the minimum context length needed for confident correct prediction.\n",
        "    - Save results and generate a histogram + power-law fit of context lengths.\n",
        "    \"\"\"\n",
        "\n",
        "    # === Clean HuggingFace cache ===\n",
        "    cache_dirs = [\n",
        "        \"/root/.cache/huggingface\",\n",
        "        \"/content/fresh_hf_cache\",\n",
        "        \"/content/hf_cache\",\n",
        "        os.path.expanduser(\"~/.cache/huggingface\"),\n",
        "        \"/tmp/huggingface_cache\",\n",
        "        \"/tmp/hf_datasets_cache\",\n",
        "        os.path.expanduser(\"~/.cache/huggingface/datasets\"),\n",
        "    ]\n",
        "    for cache_dir in cache_dirs:\n",
        "        if os.path.exists(cache_dir):\n",
        "            try:\n",
        "                shutil.rmtree(cache_dir)\n",
        "                print(f\"Removed cache directory: {cache_dir}\")\n",
        "            except:\n",
        "                pass\n",
        "\n",
        "    torch.manual_seed(42)\n",
        "\n",
        "    # === Load model ===\n",
        "    if model_name.lower() in [\"tinyllama\", \"tinyllama/tinyllama-1.1b-chat-v1.0\"]:\n",
        "        model_id = \"TinyLlama/TinyLlama-1.1B-Chat-v1.0\"\n",
        "        quant_context = nullcontext()\n",
        "        model_kwargs = {\"torch_dtype\": torch.float16}\n",
        "    else:\n",
        "        raise ValueError(f\"Unsupported model: {model_name}\")\n",
        "\n",
        "    with quant_context:\n",
        "        print(f\"Loading tokenizer for {model_id}...\")\n",
        "        tokenizer = AutoTokenizer.from_pretrained(model_id)\n",
        "        tokenizer.pad_token = tokenizer.eos_token\n",
        "        print(f\"Loading model {model_id}...\")\n",
        "        model = AutoModelForCausalLM.from_pretrained(model_id, **model_kwargs).cuda().eval()\n",
        "\n",
        "    # === Load TinyStories ===\n",
        "    print(\"\\U0001f501 Downloading TinyStories from raw URL...\")\n",
        "    url = \"https://huggingface.co/datasets/roneneldan/TinyStories/resolve/main/TinyStoriesV2-GPT4-train.txt\"\n",
        "    response = requests.get(url)\n",
        "    raw_stories = response.text.strip().split(\"\\n\\n\")\n",
        "    dataset = [{\"text\": story.strip()} for story in raw_stories if story.strip()]\n",
        "    text_key = \"text\"\n",
        "\n",
        "    # === Initialize outputs ===\n",
        "    total_correct_conf = 0\n",
        "    total_tokens = 0\n",
        "    processed = 0\n",
        "    all_ctx_lens = {}\n",
        "    conf_lens = []\n",
        "\n",
        "\n",
        "    # Analyze stories one by one, choose examples which are long enough\n",
        "    print(f\"Starting analysis of {num_stories} stories...\")\n",
        "    for ex in tqdm(dataset, desc=\"Scanning stories\"):\n",
        "        if processed >= num_stories:\n",
        "            break\n",
        "\n",
        "        text = ex[text_key]\n",
        "        tokens = tokenizer(text, return_tensors=\"pt\", truncation=False)[\"input_ids\"]\n",
        "        if tokens.shape[1] < min_tokens:\n",
        "            print(f\"Skipping story {processed}: only {tokens.shape[1]} tokens (need {min_tokens})\")\n",
        "            continue\n",
        "\n",
        "        input_ids = tokens[0][:min_tokens].cuda()\n",
        "        correct_conf, input_ids = analyze_prediction_confidence_fast(model, input_ids, conf_thresh)\n",
        "\n",
        "        valid_targets = [idx for idx in correct_conf if idx >= 32]\n",
        "        target_indices = (\n",
        "            stratified_sample(valid_targets, seq_len=len(input_ids), num_samples=num_samples_per_doc)\n",
        "            if len(valid_targets) > num_samples_per_doc else valid_targets\n",
        "        )\n",
        "\n",
        "        print(f\"Story {processed}: Found {len(correct_conf)} confident tokens, analyzing {len(target_indices)} samples\")\n",
        "\n",
        "        ctx_lens = find_tokenwise_true_context_lengths(\n",
        "            model=model,\n",
        "            tokenizer=tokenizer,\n",
        "            input_ids=input_ids,\n",
        "            target_indices=target_indices,\n",
        "            threshold=conf_thresh,\n",
        "            step_size=16,\n",
        "            use_masking=True\n",
        "        )\n",
        "\n",
        "        for tok in ctx_lens:\n",
        "            ctx_lens[tok][\"conf_ctx_len\"] = ctx_lens[tok].get(\"conf_ctx_len\")\n",
        "            if ctx_lens[tok][\"conf_ctx_len\"] is not None:\n",
        "                conf_lens.append(ctx_lens[tok][\"conf_ctx_len\"])\n",
        "\n",
        "        all_ctx_lens[f\"story_{processed}\"] = ctx_lens\n",
        "\n",
        "        os.makedirs(output_dir, exist_ok=True)\n",
        "        with open(os.path.join(output_dir, f\"context_lengths_partial.json\"), \"w\") as f:\n",
        "            json.dump(all_ctx_lens, f, indent=2)\n",
        "\n",
        "        total_correct_conf += len(correct_conf)\n",
        "        total_tokens += input_ids.shape[0] - 1\n",
        "        processed += 1\n",
        "\n",
        "        # Plot the percentage of tokens filtered to be ones which the model predicts correctly and confidently\n",
        "        plot_confidence_pie(\n",
        "            total_confident_tokens=total_correct_conf,\n",
        "            total_tokens=total_tokens,\n",
        "            processed_stories=processed,\n",
        "            save_path=os.path.join(output_dir, \"summary_classification_progress.png\")\n",
        "        )\n",
        "\n",
        "        # Plot the MCL values\n",
        "        if conf_lens:\n",
        "            plot_context_length_histogram(\n",
        "                conf_lens=conf_lens,\n",
        "                save_path=os.path.join(output_dir, \"confident_token_ctx_lengths.png\")\n",
        "            )\n",
        "\n",
        "        with open(os.path.join(output_dir, \"context_lengths.json\"), \"w\") as f:\n",
        "            json.dump(all_ctx_lens, f, indent=2)\n",
        "\n",
        "    print(f\"Analysis complete! Processed {processed} stories.\")\n",
        "    print(f\"Results saved to {output_dir}\")\n"
      ],
      "metadata": {
        "id": "U35wKaaNEsKP"
      },
      "execution_count": 5,
      "outputs": []
    },
    {
      "cell_type": "code",
      "source": [
        "\n",
        "# Run the analysis\n",
        "run_context_analysis(\n",
        "    model_name=\"tinyllama\",\n",
        "    num_stories=100,\n",
        "    min_tokens=1024,\n",
        "    conf_thresh=0.2,\n",
        "    num_samples_per_doc=100,\n",
        "    output_dir=\"./output\"\n",
        ")"
      ],
      "metadata": {
        "colab": {
          "base_uri": "https://localhost:8080/",
          "height": 1000,
          "referenced_widgets": [
            "105fb75dea9c468d98e36ba48b91d379",
            "0dbda088ec90412e959e6290392d97dc",
            "46f8cb9a4775444db4a2a7b2e0a72bf0",
            "851b3bbe139244aabede18801d4a923a",
            "21039746a8844d88883a517406891d50",
            "d3e0703dd70e486fa923ac2c37b1c879",
            "4f7ad333a1af45f48e174d01b34dc561",
            "ac84da5392dd4778b3779d85c182f357",
            "1388398ab4f74ab0831100350a1d99a7",
            "37a7646419974760845d692d0ec9ee44",
            "8dca6c4e6aee4c708b3760ab5d103c28",
            "b1678de1646e494b863cf49efa51e4b4",
            "251e1602b2a740ee8f7f8889ee2149a6",
            "c2f46f41d87b4a878dbef56baefebe84",
            "fc3c9dac87ca4c858654c1b1a50890ac",
            "d6514f03f917423fbf93a551102c28fe",
            "19c073826a644890989d6c453e561b8d",
            "63e88dbb26044265b6a75bb7112e84c7",
            "fd5c322c28e64b99a8c0e877a79e59c2",
            "8f737ff411d040dc8bf4aeb95e8c3cd6",
            "1c5b20fa7e86463db5b40a3226d28a57",
            "95568254f93e4bd29c151647bcbc36e9",
            "bfeb2fe64096492cb11f6e6548cd4085",
            "e327c7dfbb21432fb8637ca693623878",
            "c29e202a0d57417dab74485f36bf4259",
            "7a05e30b75fd499998fe2d2e1ba49de3",
            "17c3b47d4ba24f6fbddf947123b3bfbf",
            "909d46c781d64f0f85c6eb6b08a0505c",
            "f335d3386bf74e928678175451e58904",
            "000e1e07eaf44009a643eebb1bbd4333",
            "d5c3d3d16d174bf39dbffb2320b7143a",
            "78273c7bfe144d439a2a16d30652c6d5",
            "a8649eb6dcbf48d1812b4b6076a5edd3",
            "54c3b3b1aaf540568618094a4dd7c2de",
            "fc14cbf3ab3e412a9842f818596ea739",
            "6374f0370842469b95f9d316416fd71e",
            "40c8035883f44aecb7e2373861f7752e",
            "9af36ae1fa954607a5fd9207d52bae10",
            "55d85509469541e09fdb67c16196c72a",
            "f4da21a5eeef4b4bb278d9f03ee6061b",
            "cc4dc428f3734b7685e941fac87ec5ba",
            "57844d99675a49658b8e002b61e2fc12",
            "794b056f88214be3a9c662785d26d009",
            "c124d1ce85d54577be6572e2b926354f",
            "b4c10becbd7845a2828ef22975db1e74",
            "513868ac37684323b30a05bb236f3c40",
            "074bc3c411ab4be9b3101780efb62d33",
            "e085f52e520d431692e03b3d6b9890ed",
            "27578130937a49379b157b92c52aee9b",
            "aaaa468f9c174c948a6a74204af94440",
            "ebdc5604a4394bfa9bd847b583f02940",
            "50e0afcbf70b4edea8d6010587ec3bcc",
            "921ba7da88b742c19625c1aa1a32a99e",
            "7a38f868e970400b9a4a8e595cb6f08f",
            "a4e7ed98efe844a5a447faca24bbf326",
            "339f3ea0ab1b4594bd38d5e274123c1e",
            "d2c6320b1edd41829ada662c18dc5a93",
            "ce14d381e1d3436caa3a911f1f70a9d9",
            "fe880a41914943398ee851e1932c5c87",
            "931943c3c69a45969d829d89e437a937",
            "5a579c5bb74243f4b4ff0daaa3fdee8a",
            "3ac155c48bac45bba348ac0853582c24",
            "037d002dff9648cc921edd18e024bf5e",
            "c6e4a72224e0486f81763399326c0e1a",
            "89a9dfadcda34370a8b61ac2374514a8",
            "7dc638cdd4d64236aea2edd3ef286627",
            "7df660e81b844248a6deaed0decaef23",
            "c93a8faf318049f68729f2e34adef2b2",
            "2fe5bf5c529249daaea5517dc00a9c8f",
            "953fb98fe8c84f6aabba26898a98e664",
            "172d023d55814b02a96f383ab42ca5e1",
            "21c634a44479472fbeb4d20af25a35c6",
            "79a9c02dd5bf43cc92019014e1a256fd",
            "49062ef858464c9490cb50c509dc99de",
            "5f2ee295c18142a2b398afac5cd4608b",
            "40c36240b27d4cc889f312a2129021aa",
            "d4d9f8b3fcbb4490abf85183f91b2601"
          ]
        },
        "id": "_WqoGsL2uATu",
        "outputId": "e3a587d8-901f-4986-da82-dce4e03f2dae"
      },
      "execution_count": null,
      "outputs": [
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Removed cache directory: /root/.cache/huggingface\n",
            "Loading tokenizer for TinyLlama/TinyLlama-1.1B-Chat-v1.0...\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "/usr/local/lib/python3.11/dist-packages/huggingface_hub/utils/_auth.py:94: UserWarning: \n",
            "The secret `HF_TOKEN` does not exist in your Colab secrets.\n",
            "To authenticate with the Hugging Face Hub, create a token in your settings tab (https://huggingface.co/settings/tokens), set it as secret in your Google Colab and restart your session.\n",
            "You will be able to reuse this secret in all of your notebooks.\n",
            "Please note that authentication is recommended but still optional to access public models or datasets.\n",
            "  warnings.warn(\n"
          ]
        },
        {
          "output_type": "display_data",
          "data": {
            "text/plain": [
              "tokenizer_config.json:   0%|          | 0.00/1.29k [00:00<?, ?B/s]"
            ],
            "application/vnd.jupyter.widget-view+json": {
              "version_major": 2,
              "version_minor": 0,
              "model_id": "105fb75dea9c468d98e36ba48b91d379"
            }
          },
          "metadata": {}
        },
        {
          "output_type": "display_data",
          "data": {
            "text/plain": [
              "tokenizer.model:   0%|          | 0.00/500k [00:00<?, ?B/s]"
            ],
            "application/vnd.jupyter.widget-view+json": {
              "version_major": 2,
              "version_minor": 0,
              "model_id": "b1678de1646e494b863cf49efa51e4b4"
            }
          },
          "metadata": {}
        },
        {
          "output_type": "display_data",
          "data": {
            "text/plain": [
              "tokenizer.json:   0%|          | 0.00/1.84M [00:00<?, ?B/s]"
            ],
            "application/vnd.jupyter.widget-view+json": {
              "version_major": 2,
              "version_minor": 0,
              "model_id": "bfeb2fe64096492cb11f6e6548cd4085"
            }
          },
          "metadata": {}
        },
        {
          "output_type": "display_data",
          "data": {
            "text/plain": [
              "special_tokens_map.json:   0%|          | 0.00/551 [00:00<?, ?B/s]"
            ],
            "application/vnd.jupyter.widget-view+json": {
              "version_major": 2,
              "version_minor": 0,
              "model_id": "54c3b3b1aaf540568618094a4dd7c2de"
            }
          },
          "metadata": {}
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Loading model TinyLlama/TinyLlama-1.1B-Chat-v1.0...\n"
          ]
        },
        {
          "output_type": "display_data",
          "data": {
            "text/plain": [
              "config.json:   0%|          | 0.00/608 [00:00<?, ?B/s]"
            ],
            "application/vnd.jupyter.widget-view+json": {
              "version_major": 2,
              "version_minor": 0,
              "model_id": "b4c10becbd7845a2828ef22975db1e74"
            }
          },
          "metadata": {}
        },
        {
          "output_type": "display_data",
          "data": {
            "text/plain": [
              "model.safetensors:   0%|          | 0.00/2.20G [00:00<?, ?B/s]"
            ],
            "application/vnd.jupyter.widget-view+json": {
              "version_major": 2,
              "version_minor": 0,
              "model_id": "339f3ea0ab1b4594bd38d5e274123c1e"
            }
          },
          "metadata": {}
        },
        {
          "output_type": "display_data",
          "data": {
            "text/plain": [
              "generation_config.json:   0%|          | 0.00/124 [00:00<?, ?B/s]"
            ],
            "application/vnd.jupyter.widget-view+json": {
              "version_major": 2,
              "version_minor": 0,
              "model_id": "7df660e81b844248a6deaed0decaef23"
            }
          },
          "metadata": {}
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "🔁 Downloading TinyStories from raw URL...\n",
            "Starting analysis of 100 stories...\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\rScanning stories:   0%|          | 0/739990 [00:00<?, ?it/s]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 0: only 553 tokens (need 1024)\n",
            "Skipping story 0: only 934 tokens (need 1024)\n",
            "Skipping story 0: only 529 tokens (need 1024)\n",
            "Skipping story 0: only 342 tokens (need 1024)\n",
            "Story 0: Found 519 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   1%|          | 1/100 [00:00<00:13,  7.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:07, 13.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:04, 18.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:03, 24.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:03, 25.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:03, 24.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:00<00:03, 24.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:00<00:03, 25.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:02, 25.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:02, 24.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:01<00:03, 22.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:01<00:03, 21.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:06, 10.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:10,  5.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:17,  3.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:04<00:13,  4.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:04<00:10,  5.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:05<00:16,  3.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:06<00:12,  4.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:06<00:09,  5.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:06<00:08,  5.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:06,  6.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:06<00:05,  7.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:09<00:18,  2.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:09<00:13,  2.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:09<00:09,  3.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:07,  4.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:06,  5.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:10<00:05,  5.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:10<00:06,  4.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:10<00:05,  5.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:10<00:05,  5.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:06,  4.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:11<00:05,  5.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:11<00:04,  5.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:06,  3.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:12<00:07,  3.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:12<00:06,  3.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:12<00:05,  4.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:13<00:08,  2.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:13<00:06,  3.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:13<00:04,  4.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:13<00:04,  4.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:13<00:03,  5.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:13<00:02,  5.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:13<00:02,  6.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:13<00:02,  6.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:14<00:02,  6.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:14<00:01,  7.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:15<00:06,  1.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:15<00:04,  2.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:15<00:03,  3.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:16<00:03,  2.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:16<00:02,  3.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:16<00:01,  3.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:18<00:04,  1.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:18<00:02,  1.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:18<00:01,  2.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:18<00:01,  2.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:19<00:00,  3.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:19<00:00,  3.97it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:19<00:00,  5.14it/s]\n",
            "Scanning stories:   0%|          | 5/739990 [00:21<879:33:30,  4.28s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 1: only 184 tokens (need 1024)\n",
            "Skipping story 1: only 980 tokens (need 1024)\n",
            "Skipping story 1: only 421 tokens (need 1024)\n",
            "Skipping story 1: only 320 tokens (need 1024)\n",
            "Skipping story 1: only 929 tokens (need 1024)\n",
            "Skipping story 1: only 146 tokens (need 1024)\n",
            "Skipping story 1: only 459 tokens (need 1024)\n",
            "Skipping story 1: only 292 tokens (need 1024)\n",
            "Story 1: Found 550 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:02, 31.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:02, 31.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:05, 15.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:05, 13.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:05, 14.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:04, 16.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:04, 17.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:04, 14.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:05, 12.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:03<00:18,  3.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:03<00:13,  4.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:03<00:10,  5.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:04<00:08,  6.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:04<00:07,  8.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:04<00:07,  7.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:04<00:06,  8.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:04<00:05,  9.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:05<00:05,  9.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:05<00:04, 10.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:07,  6.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:06<00:06,  7.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:06<00:09,  4.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:07<00:08,  4.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:07<00:06,  5.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:05,  6.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:07<00:04,  7.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:07<00:04,  7.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:07<00:04,  7.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:08<00:04,  7.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:08<00:04,  7.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:08<00:03,  7.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:08<00:03,  7.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:08<00:04,  6.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:08<00:04,  6.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:12<00:29,  1.13s/it]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:12<00:21,  1.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:12<00:15,  1.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:12<00:11,  2.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:12<00:09,  2.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:13<00:07,  2.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:13<00:05,  3.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:15<00:16,  1.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:16<00:17,  1.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:16<00:12,  1.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:17<00:08,  1.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:17<00:06,  2.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:17<00:07,  1.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:20<00:14,  1.10s/it]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:21<00:13,  1.11s/it]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:21<00:09,  1.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:23<00:10,  1.07s/it]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:23<00:07,  1.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:23<00:04,  1.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:23<00:03,  2.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:24<00:02,  2.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:24<00:01,  2.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:24<00:01,  3.36it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:24<00:00,  3.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:24<00:00,  4.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:25<00:00,  4.73it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:27<00:00,  3.67it/s]\n",
            "Scanning stories:   0%|          | 14/739990 [00:49<705:36:22,  3.43s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 2: only 668 tokens (need 1024)\n",
            "Skipping story 2: only 223 tokens (need 1024)\n",
            "Skipping story 2: only 645 tokens (need 1024)\n",
            "Story 2: Found 515 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:04, 23.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:   7%|▋         | 7/100 [00:00<00:03, 28.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:03, 26.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:04, 19.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:04, 16.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:01<00:05, 14.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:05, 14.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:04, 17.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:03, 19.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:07,  9.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:07,  9.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:06, 10.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:06, 10.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:05, 11.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:02<00:04, 12.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:02<00:04, 12.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:09,  6.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:04<00:10,  5.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:05<00:15,  3.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:06<00:19,  2.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:06<00:14,  3.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:06<00:10,  4.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:08,  5.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:07<00:06,  6.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:07<00:05,  7.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:07<00:05,  7.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:07<00:04,  7.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:04,  8.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:12,  2.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:10,  3.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:09<00:08,  3.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:09<00:07,  4.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:09<00:06,  5.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:09<00:06,  4.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:05,  5.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:10<00:07,  3.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:10<00:06,  4.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:10<00:05,  5.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:10<00:04,  5.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:11<00:04,  5.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:11<00:04,  4.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:11<00:04,  5.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:12<00:06,  3.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:12<00:05,  3.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:12<00:04,  4.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:12<00:03,  5.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:12<00:02,  5.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:12<00:02,  6.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:12<00:02,  6.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:13<00:02,  6.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:13<00:01,  6.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:15<00:08,  1.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:15<00:06,  1.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:15<00:04,  2.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:16<00:04,  2.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:16<00:03,  2.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:16<00:02,  2.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:16<00:01,  3.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:17<00:01,  2.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:17<00:01,  2.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:18<00:01,  2.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:18<00:00,  3.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:19<00:00,  1.57it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:20<00:00,  4.99it/s]\n",
            "Scanning stories:   0%|          | 18/739990 [01:10<817:53:37,  3.98s/it]Token indices sequence length is longer than the specified maximum sequence length for this model (2210 > 2048). Running this sequence through the model will result in indexing errors\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 3: only 459 tokens (need 1024)\n",
            "Skipping story 3: only 175 tokens (need 1024)\n",
            "Story 3: Found 539 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 33.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:03, 28.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:03, 26.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:06, 13.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:01<00:07, 11.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:05, 14.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:06, 12.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:04, 15.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:02<00:08,  8.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:02<00:14,  5.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:03<00:11,  6.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:03<00:09,  6.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:03<00:08,  7.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:03<00:08,  7.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:04<00:15,  3.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:04<00:11,  5.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:05<00:13,  4.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:05<00:12,  4.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:06<00:12,  4.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:06<00:14,  3.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:06<00:13,  3.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:06<00:09,  5.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:07<00:07,  6.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:09<00:25,  1.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:09<00:17,  2.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:10<00:12,  3.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:10<00:09,  4.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:10<00:10,  3.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:11<00:09,  4.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:11<00:07,  4.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:11<00:06,  5.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:11<00:05,  5.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:11<00:05,  6.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:11<00:04,  7.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:11<00:04,  7.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:12<00:05,  5.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:12<00:05,  5.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:12<00:04,  6.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:12<00:04,  6.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:12<00:06,  4.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:13<00:05,  4.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:13<00:04,  5.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:13<00:05,  4.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:14<00:08,  2.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:14<00:06,  3.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:14<00:04,  4.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:14<00:04,  4.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:14<00:03,  5.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:14<00:02,  6.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:17<00:16,  1.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:18<00:15,  1.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:18<00:11,  1.36it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:18<00:07,  1.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:19<00:05,  2.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:19<00:04,  2.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:19<00:03,  3.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:23<00:14,  1.46s/it]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:26<00:16,  1.83s/it]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:26<00:10,  1.37s/it]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:26<00:07,  1.01s/it]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:28<00:07,  1.31s/it]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:28<00:04,  1.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:29<00:02,  1.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:29<00:01,  1.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:29<00:00,  2.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:29<00:00,  2.87it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:29<00:00,  3.36it/s]\n",
            "Scanning stories:   0%|          | 21/739990 [01:40<1136:23:36,  5.53s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 4: only 455 tokens (need 1024)\n",
            "Skipping story 4: only 177 tokens (need 1024)\n",
            "Skipping story 4: only 638 tokens (need 1024)\n",
            "Story 4: Found 533 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 29.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:03, 24.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:04, 20.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:04, 21.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:06, 13.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:04, 16.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:04, 18.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:04, 15.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:06, 11.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:01<00:05, 12.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:06, 11.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:05, 12.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:05, 12.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:07,  8.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:02<00:06, 10.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:03<00:07,  8.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:06,  9.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:03<00:06,  8.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:03<00:05,  9.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:04<00:04, 10.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:04<00:04, 10.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:04<00:04, 11.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:04<00:05,  8.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:04<00:05,  7.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:05<00:09,  4.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:05<00:07,  5.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:05<00:05,  6.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:05<00:05,  7.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:06<00:04,  7.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:06<00:04,  7.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:06<00:04,  8.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:06<00:04,  8.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:06<00:04,  6.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:07<00:07,  4.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:07<00:06,  4.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:07<00:06,  4.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:07<00:06,  4.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:07<00:05,  5.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:07<00:04,  6.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:08<00:04,  6.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:10<00:18,  1.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:10<00:13,  1.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:10<00:09,  2.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:10<00:07,  3.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:10<00:07,  2.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:13<00:22,  1.11s/it]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:13<00:15,  1.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:14<00:11,  1.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:14<00:08,  2.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:14<00:05,  2.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:14<00:04,  3.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:14<00:03,  3.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:15<00:04,  2.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:16<00:06,  1.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:16<00:05,  1.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:16<00:04,  2.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:16<00:02,  3.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:20<00:09,  1.25s/it]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:20<00:06,  1.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:22<00:07,  1.23s/it]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:22<00:04,  1.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:22<00:02,  1.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:22<00:01,  1.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:23<00:00,  2.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:23<00:00,  2.92it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:23<00:00,  4.27it/s]\n",
            "Scanning stories:   0%|          | 25/739990 [02:04<1168:36:49,  5.69s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 5: only 489 tokens (need 1024)\n",
            "Story 5: Found 563 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:04, 18.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:03, 23.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:04, 20.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:01<00:05, 15.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:04, 17.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:04, 17.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:03, 18.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:04, 14.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:04, 15.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:01<00:04, 16.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:07,  8.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:07,  7.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:02<00:06,  9.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:03<00:06,  9.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:03<00:06,  8.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:03<00:07,  7.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:04<00:11,  4.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:04<00:11,  4.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:05<00:09,  5.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:05<00:11,  4.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:05<00:11,  4.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:07<00:26,  1.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:08<00:17,  2.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:08<00:21,  2.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:09<00:14,  2.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:09<00:10,  3.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:09<00:08,  4.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:09<00:07,  4.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:09<00:06,  5.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:05,  5.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:05,  6.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:10<00:04,  7.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:10<00:07,  4.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:10<00:06,  5.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:10<00:05,  5.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:04,  6.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:10<00:04,  6.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:11<00:03,  7.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:03,  7.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:11<00:03,  7.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:11<00:03,  7.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:11<00:02,  7.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:13<00:12,  1.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:14<00:17,  1.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:14<00:12,  1.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:14<00:09,  2.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:15<00:06,  2.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:15<00:05,  3.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:15<00:04,  3.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:15<00:03,  4.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:15<00:02,  5.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:15<00:02,  4.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:16<00:02,  5.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:16<00:02,  5.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:16<00:01,  5.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:17<00:04,  1.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:20<00:10,  1.32s/it]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:21<00:06,  1.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:21<00:04,  1.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:21<00:03,  1.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:21<00:01,  2.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:22<00:01,  1.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:22<00:00,  2.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:22<00:00,  2.98it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:22<00:00,  4.39it/s]\n",
            "Scanning stories:   0%|          | 27/739990 [02:27<1401:35:51,  6.82s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 6: only 855 tokens (need 1024)\n",
            "Skipping story 6: only 585 tokens (need 1024)\n",
            "Skipping story 6: only 173 tokens (need 1024)\n",
            "Skipping story 6: only 194 tokens (need 1024)\n",
            "Story 6: Found 531 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:03, 27.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:   7%|▋         | 7/100 [00:00<00:03, 25.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:04, 20.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:04, 20.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:00<00:03, 23.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:04, 17.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:05, 13.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:05, 14.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:04, 15.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:04, 14.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:02<00:10,  6.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:03<00:15,  4.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:03<00:13,  4.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:03<00:10,  5.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:04<00:08,  7.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:04<00:07,  8.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:04<00:10,  5.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:05<00:08,  6.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:05<00:06,  7.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:05<00:05,  8.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:05<00:07,  6.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:06<00:12,  3.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:09,  4.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:07<00:08,  5.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:07<00:06,  6.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:07<00:06,  6.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:09<00:15,  2.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:09<00:10,  3.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:09,  3.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:07,  4.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:09<00:06,  4.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:09<00:05,  5.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:09<00:04,  6.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:14<00:42,  1.41s/it]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:14<00:30,  1.05s/it]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:15<00:22,  1.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:15<00:18,  1.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:15<00:13,  1.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:15<00:10,  2.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:15<00:07,  3.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:16<00:07,  2.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:16<00:05,  3.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:16<00:04,  4.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:16<00:03,  5.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:16<00:04,  3.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:17<00:04,  3.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:17<00:03,  4.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:17<00:03,  5.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:17<00:02,  5.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:21<00:17,  1.27s/it]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:21<00:13,  1.02s/it]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:21<00:09,  1.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:22<00:06,  1.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:23<00:07,  1.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:23<00:04,  1.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:23<00:03,  2.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:23<00:02,  2.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:29<00:11,  1.89s/it]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:29<00:06,  1.37s/it]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:31<00:06,  1.57s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:31<00:03,  1.15s/it]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:31<00:01,  1.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:31<00:00,  1.55it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:32<00:00,  3.10it/s]\n",
            "Scanning stories:   0%|          | 32/739990 [03:00<1382:56:42,  6.73s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 7: only 886 tokens (need 1024)\n",
            "Skipping story 7: only 820 tokens (need 1024)\n",
            "Skipping story 7: only 191 tokens (need 1024)\n",
            "Skipping story 7: only 1009 tokens (need 1024)\n",
            "Story 7: Found 525 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:03, 26.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:03, 29.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:02, 30.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:00<00:03, 23.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:00<00:03, 21.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:05, 13.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:07,  9.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:02<00:08,  8.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:02<00:09,  7.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:10,  6.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:10,  6.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:03<00:07,  8.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:03<00:09,  6.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:03<00:09,  6.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:10,  5.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:05<00:17,  3.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:05<00:15,  3.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:05<00:10,  4.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:05<00:08,  5.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:06<00:13,  3.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:07<00:14,  3.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:07<00:10,  4.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:07<00:07,  5.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:07<00:07,  5.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:07<00:06,  6.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:07<00:05,  7.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:08<00:05,  7.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:08<00:05,  7.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:08<00:07,  4.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:08<00:06,  5.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:09,  3.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:10<00:14,  2.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:10<00:18,  1.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:11<00:14,  2.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:11<00:11,  2.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:12<00:18,  1.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:14<00:29,  1.02s/it]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:14<00:21,  1.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:15<00:18,  1.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:15<00:15,  1.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:15<00:11,  2.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:15<00:08,  2.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:16<00:08,  2.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:16<00:06,  3.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:16<00:05,  4.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:16<00:05,  3.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:17<00:07,  2.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:20<00:19,  1.11s/it]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:20<00:13,  1.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:20<00:09,  1.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:20<00:06,  2.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:20<00:05,  2.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:20<00:03,  3.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:22<00:08,  1.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:22<00:05,  1.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:22<00:03,  2.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:22<00:02,  3.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:23<00:02,  3.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:23<00:01,  4.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:23<00:01,  3.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:23<00:01,  4.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:23<00:00,  4.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:24<00:01,  2.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:24<00:00,  3.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:28<00:01,  1.19s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:28<00:00,  3.54it/s]\n",
            "Scanning stories:   0%|          | 37/739990 [03:29<1312:42:10,  6.39s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 8: only 898 tokens (need 1024)\n",
            "Skipping story 8: only 270 tokens (need 1024)\n",
            "Skipping story 8: only 601 tokens (need 1024)\n",
            "Skipping story 8: only 480 tokens (need 1024)\n",
            "Skipping story 8: only 458 tokens (need 1024)\n",
            "Story 8: Found 534 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 28.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:   7%|▋         | 7/100 [00:00<00:03, 30.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:03, 23.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:03, 23.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:00<00:03, 23.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:00<00:03, 24.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:00<00:03, 24.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:06, 12.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:06, 11.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:06, 10.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:02<00:06, 11.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:05, 12.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:05, 12.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:04, 13.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:02<00:04, 13.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:02<00:05, 10.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:02<00:04, 11.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:03<00:04, 12.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:03<00:04, 12.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:03<00:03, 12.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:03<00:05,  9.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:03<00:04, 10.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:04<00:06,  6.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:05<00:13,  3.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:05<00:09,  4.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:05<00:07,  5.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:06<00:06,  6.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:14,  2.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:07<00:12,  2.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:07<00:10,  3.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:08<00:08,  3.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:08<00:07,  4.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:08<00:06,  4.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:08<00:05,  5.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:08<00:04,  6.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:08<00:04,  6.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:09<00:05,  4.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:09<00:04,  5.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:09<00:06,  3.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:09<00:05,  4.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:10<00:05,  4.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:10<00:06,  3.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:10<00:04,  4.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:13<00:22,  1.14s/it]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:13<00:15,  1.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:14<00:11,  1.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:14<00:08,  2.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:14<00:06,  2.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:14<00:04,  3.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:14<00:03,  3.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:14<00:03,  3.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:15<00:02,  4.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:15<00:02,  4.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:15<00:02,  3.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:15<00:02,  4.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:16<00:01,  4.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:16<00:01,  4.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:16<00:01,  5.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:16<00:00,  5.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:16<00:00,  5.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:16<00:00,  4.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:17<00:00,  4.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:17<00:00,  5.25it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:21<00:00,  4.62it/s]\n",
            "Scanning stories:   0%|          | 43/739990 [03:52<1102:42:01,  5.36s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 9: only 802 tokens (need 1024)\n",
            "Story 9: Found 506 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:04, 22.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:05, 17.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:05, 16.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:01<00:06, 12.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:05, 14.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:04, 16.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:04, 17.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:04, 16.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:01<00:04, 16.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:01<00:03, 17.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:01<00:03, 18.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:03, 17.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:04, 14.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:02<00:04, 15.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:02<00:05, 11.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:11,  4.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:03<00:09,  5.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:04<00:07,  6.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:04<00:06,  7.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:04<00:05,  8.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:04<00:05,  8.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:04<00:04,  9.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:04<00:04, 10.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:06<00:14,  2.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:06<00:13,  3.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:07<00:17,  2.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:08<00:15,  2.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:08<00:13,  2.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:08<00:10,  3.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:08,  3.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:09<00:07,  4.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:09<00:07,  4.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:09<00:05,  5.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:09<00:05,  5.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:09<00:04,  6.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:09<00:04,  6.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:10<00:08,  3.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:09,  2.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:11<00:07,  3.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:12<00:11,  2.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:12<00:08,  2.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:12<00:06,  3.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:12<00:05,  4.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:12<00:04,  4.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:13<00:10,  1.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:13<00:07,  2.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:14<00:06,  2.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:14<00:05,  3.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:14<00:03,  3.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:14<00:03,  4.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:14<00:02,  5.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:15<00:02,  5.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:15<00:01,  6.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:15<00:01,  6.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:15<00:01,  6.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:15<00:01,  6.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:15<00:01,  6.36it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:16<00:01,  4.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:16<00:01,  4.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:19<00:04,  1.13s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:20<00:03,  1.18s/it]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:21<00:01,  1.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:21<00:00,  1.52it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:21<00:00,  4.67it/s]\n",
            "Scanning stories:   0%|          | 45/739990 [04:14<1283:22:50,  6.24s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 10: only 367 tokens (need 1024)\n",
            "Story 10: Found 534 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:03, 29.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:04, 19.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:05, 16.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:00<00:04, 19.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:06, 12.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:05, 14.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:02<00:08,  8.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:02<00:07, 10.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:02<00:07,  9.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:02<00:07,  9.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:06, 10.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:05, 12.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:04, 13.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:03<00:04, 13.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:03<00:04, 12.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:03<00:04, 12.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:03<00:04, 12.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:03<00:04, 12.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:04<00:05,  9.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:04<00:04,  9.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:04<00:04, 10.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:05<00:13,  3.36it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:06<00:09,  4.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:06<00:07,  5.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:06<00:07,  5.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:07,  5.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:07<00:05,  5.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:07<00:05,  5.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:07<00:05,  6.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:08<00:07,  4.36it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:08<00:06,  4.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:09<00:11,  2.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:09<00:08,  3.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:09<00:07,  3.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:09<00:07,  3.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:10<00:07,  3.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:10<00:06,  3.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:10<00:05,  4.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:10<00:05,  3.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:10<00:04,  4.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:11<00:03,  5.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:11<00:03,  5.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:11<00:06,  2.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:12<00:04,  3.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:12<00:03,  4.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:12<00:04,  3.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:12<00:03,  4.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:13<00:04,  2.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:13<00:03,  3.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:13<00:02,  4.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:13<00:02,  4.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:14<00:02,  3.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:14<00:02,  3.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:14<00:01,  4.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:14<00:01,  4.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:18<00:06,  1.33s/it]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:22<00:08,  2.18s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:25<00:07,  2.40s/it]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:25<00:03,  1.73s/it]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:26<00:01,  1.26s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:26<00:00,  3.82it/s]\n",
            "Scanning stories:   0%|          | 47/739990 [04:41<1552:08:39,  7.55s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 11: only 406 tokens (need 1024)\n",
            "Story 11: Found 562 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:03, 31.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:02, 30.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:03, 23.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:03, 25.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:00<00:05, 16.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:04, 17.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:04, 18.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:05, 12.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:01<00:05, 13.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:01<00:05, 13.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:01<00:04, 14.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:10,  6.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:08,  7.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:08,  6.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:12,  4.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:04<00:10,  5.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:04<00:08,  6.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:04<00:07,  6.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:04<00:06,  7.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:04<00:05,  8.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:05<00:06,  7.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:15,  2.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:07<00:16,  2.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:07<00:11,  3.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:10<00:23,  1.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:10<00:17,  2.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:11<00:17,  2.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:11<00:19,  1.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:11<00:15,  2.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:12<00:12,  2.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:12<00:09,  3.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:12<00:07,  4.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:12<00:06,  4.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:12<00:05,  5.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:12<00:04,  6.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:12<00:04,  6.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:12<00:03,  6.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:13<00:03,  7.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:13<00:06,  3.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:13<00:05,  4.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:13<00:04,  5.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:14<00:03,  5.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:14<00:03,  6.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:14<00:03,  6.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:14<00:02,  6.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:14<00:03,  5.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:15<00:04,  3.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:15<00:03,  4.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:15<00:03,  4.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:15<00:02,  5.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:17<00:07,  1.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:17<00:05,  2.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:17<00:03,  2.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:17<00:02,  3.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:17<00:02,  4.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:17<00:01,  4.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:18<00:01,  4.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:18<00:01,  4.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:18<00:01,  4.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:18<00:00,  4.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:18<00:00,  4.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:19<00:00,  2.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:20<00:00,  2.77it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:20<00:00,  4.95it/s]\n",
            "Scanning stories:   0%|          | 49/739990 [05:02<1674:12:09,  8.15s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 12: only 263 tokens (need 1024)\n",
            "Skipping story 12: only 248 tokens (need 1024)\n",
            "Skipping story 12: only 500 tokens (need 1024)\n",
            "Skipping story 12: only 239 tokens (need 1024)\n",
            "Story 12: Found 564 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 25.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:   7%|▋         | 7/100 [00:00<00:03, 26.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:03, 24.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:03, 24.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:03, 23.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:00<00:04, 16.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:04, 18.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:04, 17.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:03, 18.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:03, 18.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:03, 18.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:01<00:03, 18.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:05, 10.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:05, 11.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:02<00:05, 11.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:02<00:04, 11.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:02<00:04, 12.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:02<00:04, 11.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:03<00:05,  9.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:03<00:06,  8.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:03<00:07,  6.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:04<00:06,  7.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:04<00:05,  8.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:04<00:06,  6.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:04<00:05,  7.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:04<00:05,  7.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:05<00:04,  8.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:05<00:04,  8.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:05<00:04,  8.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:05<00:04,  8.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:05<00:04,  8.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:05<00:03,  8.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:05<00:03,  8.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:05<00:03,  8.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:05<00:03,  8.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:06<00:03,  8.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:06<00:03,  8.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:06<00:03,  8.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:06<00:03,  8.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:06<00:04,  5.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:09<00:25,  1.00s/it]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:09<00:17,  1.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:09<00:12,  1.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:10<00:09,  2.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:14<00:35,  1.68s/it]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:14<00:24,  1.22s/it]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:16<00:23,  1.23s/it]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:16<00:16,  1.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:16<00:11,  1.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:16<00:08,  1.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:16<00:05,  2.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:16<00:04,  3.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:16<00:03,  3.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:17<00:02,  4.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:17<00:02,  4.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:17<00:01,  5.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:19<00:07,  1.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:20<00:06,  1.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:20<00:04,  1.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:20<00:02,  2.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:20<00:01,  2.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:21<00:01,  3.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:21<00:00,  3.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:22<00:00,  2.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:22<00:00,  2.84it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:22<00:00,  4.42it/s]\n",
            "Scanning stories:   0%|          | 54/739990 [05:25<1370:49:15,  6.67s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 13: only 258 tokens (need 1024)\n",
            "Story 13: Found 552 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 29.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:03, 26.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:03, 28.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:02, 30.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:00<00:04, 17.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:06, 12.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:08,  9.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:07, 10.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:05, 12.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:05, 13.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:05, 12.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:05, 11.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:05, 12.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:02<00:04, 13.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:02<00:04, 13.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:04, 13.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:03<00:04, 13.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:03<00:03, 13.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:03<00:04, 11.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:03<00:04, 10.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:04<00:04, 10.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:04<00:04, 11.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:04<00:07,  5.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:05<00:07,  5.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:05<00:05,  6.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:05<00:09,  4.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:06<00:07,  4.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:06<00:06,  5.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:06<00:06,  5.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:06<00:06,  5.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:06<00:05,  6.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:06<00:04,  6.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:07<00:04,  7.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:07<00:10,  2.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:08<00:09,  2.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:08<00:07,  3.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:08<00:08,  3.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:08<00:06,  4.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:08<00:05,  4.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:09<00:04,  5.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:09<00:03,  5.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:11<00:18,  1.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:11<00:13,  1.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:11<00:09,  2.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:12<00:07,  2.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:12<00:05,  3.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:12<00:04,  3.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:14<00:11,  1.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:14<00:08,  1.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:15<00:11,  1.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:15<00:08,  1.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:16<00:08,  1.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:17<00:06,  1.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:17<00:04,  2.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:17<00:03,  2.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:17<00:02,  3.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:17<00:01,  3.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:18<00:01,  4.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:18<00:01,  4.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:18<00:00,  4.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:18<00:00,  4.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:18<00:00,  4.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:19<00:00,  5.00it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:19<00:00,  5.22it/s]\n",
            "Scanning stories:   0%|          | 56/739990 [05:45<1501:41:53,  7.31s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 14: only 234 tokens (need 1024)\n",
            "Skipping story 14: only 419 tokens (need 1024)\n",
            "Skipping story 14: only 1018 tokens (need 1024)\n",
            "Skipping story 14: only 397 tokens (need 1024)\n",
            "Story 14: Found 529 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:03, 26.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:   7%|▋         | 7/100 [00:00<00:04, 22.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:05, 17.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:03, 22.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:00<00:05, 14.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:04, 18.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:03, 19.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:03, 20.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:04, 14.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:04, 14.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:05, 13.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:09,  6.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:03<00:08,  7.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:03<00:10,  5.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:14,  3.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:04<00:11,  5.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:04<00:08,  6.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:04<00:08,  6.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:05<00:07,  7.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:05<00:07,  6.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:05<00:05,  7.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:06,  6.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:05<00:05,  7.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:06<00:04,  8.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:06<00:04,  9.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:07<00:09,  3.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:07<00:08,  4.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:07,  4.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:07<00:06,  5.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:07<00:05,  6.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:07<00:04,  6.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:08<00:04,  7.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:08<00:07,  4.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:08<00:05,  5.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:11<00:26,  1.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:11<00:18,  1.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:11<00:13,  1.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:10,  2.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:11<00:07,  3.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:11<00:06,  3.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:12<00:04,  4.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:12<00:04,  5.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:12<00:03,  5.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:12<00:03,  6.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:12<00:02,  6.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:12<00:02,  6.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:12<00:02,  7.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:13<00:02,  5.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:13<00:02,  6.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:13<00:02,  6.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:13<00:01,  6.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:13<00:01,  6.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:13<00:01,  6.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:13<00:01,  6.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:14<00:01,  6.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:15<00:04,  1.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:15<00:03,  2.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:15<00:02,  2.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:16<00:01,  3.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:16<00:01,  2.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:16<00:00,  3.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:16<00:00,  3.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:17<00:00,  4.34it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:17<00:00,  5.78it/s]\n",
            "Scanning stories:   0%|          | 61/739990 [06:03<1189:19:49,  5.79s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 15: Found 524 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:03, 31.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:04, 20.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:04, 21.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:03, 25.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:00<00:04, 19.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:00<00:03, 21.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:04, 18.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:03, 19.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:03, 17.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:03, 17.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:01<00:04, 14.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:01<00:04, 14.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:04, 14.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:02<00:04, 14.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:02<00:04, 12.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:02<00:04, 12.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:02<00:04, 13.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:03<00:05,  9.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:03<00:07,  6.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:03<00:06,  8.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:03<00:05,  8.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:04<00:04,  9.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:04<00:04, 10.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:04<00:03, 10.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:04<00:05,  7.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:04<00:04,  7.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:05<00:04,  7.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:05<00:04,  8.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:05<00:06,  5.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:05<00:07,  4.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:05<00:06,  5.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:06<00:07,  4.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:06<00:10,  3.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:07<00:07,  3.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:07<00:06,  4.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:07<00:05,  5.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:07<00:04,  5.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:07<00:06,  3.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:08<00:05,  4.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:10<00:17,  1.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:10<00:12,  1.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:10<00:09,  2.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:10<00:06,  2.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:10<00:05,  3.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:10<00:04,  4.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:10<00:03,  4.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:10<00:02,  5.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:11<00:02,  5.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:11<00:02,  6.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:12<00:06,  2.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:12<00:05,  2.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:13<00:04,  2.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:13<00:03,  3.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:18<00:17,  1.89s/it]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:19<00:12,  1.51s/it]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:19<00:07,  1.11s/it]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:19<00:04,  1.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:19<00:03,  1.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:21<00:03,  1.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:21<00:02,  1.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:21<00:01,  1.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:22<00:00,  2.29it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:22<00:00,  4.39it/s]\n",
            "Scanning stories:   0%|          | 62/739990 [06:27<1581:33:12,  7.69s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 16: only 688 tokens (need 1024)\n",
            "Skipping story 16: only 194 tokens (need 1024)\n",
            "Skipping story 16: only 620 tokens (need 1024)\n",
            "Skipping story 16: only 563 tokens (need 1024)\n",
            "Skipping story 16: only 535 tokens (need 1024)\n",
            "Skipping story 16: only 344 tokens (need 1024)\n",
            "Skipping story 16: only 615 tokens (need 1024)\n",
            "Skipping story 16: only 865 tokens (need 1024)\n",
            "Skipping story 16: only 520 tokens (need 1024)\n",
            "Skipping story 16: only 226 tokens (need 1024)\n",
            "Skipping story 16: only 920 tokens (need 1024)\n",
            "Skipping story 16: only 421 tokens (need 1024)\n",
            "Skipping story 16: only 973 tokens (need 1024)\n",
            "Skipping story 16: only 498 tokens (need 1024)\n",
            "Skipping story 16: only 481 tokens (need 1024)\n",
            "Skipping story 16: only 237 tokens (need 1024)\n",
            "Skipping story 16: only 570 tokens (need 1024)\n",
            "Skipping story 16: only 131 tokens (need 1024)\n",
            "Skipping story 16: only 394 tokens (need 1024)\n",
            "Story 16: Found 556 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 28.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:03, 27.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:03, 22.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:04, 20.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:05, 15.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:01<00:06, 13.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:05, 13.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:05, 14.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:05, 13.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:04, 15.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:01<00:05, 13.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:05, 13.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:04, 14.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:03<00:13,  4.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:03<00:10,  5.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:09,  6.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:15,  3.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:04<00:14,  4.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:05<00:16,  3.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:06<00:26,  2.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:06<00:21,  2.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:06<00:18,  2.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:06<00:12,  4.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:07<00:08,  5.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:07<00:06,  7.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:07<00:05,  8.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:07<00:04,  9.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:07<00:04,  9.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:07<00:03, 10.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:08<00:07,  5.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:08<00:06,  5.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:08<00:05,  6.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:06,  5.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:05,  6.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:09<00:04,  6.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:09<00:05,  5.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:09<00:04,  6.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:09<00:05,  5.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:04,  6.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:10<00:06,  4.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:10<00:05,  5.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:10<00:04,  5.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:10<00:03,  6.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:10<00:03,  6.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:11<00:03,  7.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:11<00:02,  7.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:11<00:02,  7.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:11<00:02,  7.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:11<00:03,  5.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:11<00:02,  6.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:12<00:03,  4.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:12<00:05,  3.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:12<00:03,  3.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:13<00:03,  4.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:13<00:02,  5.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:13<00:03,  3.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:13<00:02,  3.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:14<00:02,  4.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:14<00:01,  4.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:14<00:01,  5.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:14<00:01,  5.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:14<00:01,  5.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:14<00:00,  5.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:16<00:02,  1.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:16<00:01,  2.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:16<00:00,  2.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:16<00:00,  3.12it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:17<00:00,  5.84it/s]\n",
            "Scanning stories:   0%|          | 82/739990 [06:45<533:10:48,  2.59s/it] "
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 17: only 339 tokens (need 1024)\n",
            "Skipping story 17: only 300 tokens (need 1024)\n",
            "Story 17: Found 571 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:03, 25.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:03, 29.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:04, 21.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:03, 21.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:00<00:04, 18.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:05, 15.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:04, 17.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:03, 18.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:03, 18.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:01<00:04, 15.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:05, 11.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:09,  6.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:04<00:17,  3.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:13,  4.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:04<00:10,  5.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:05<00:11,  4.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:05<00:08,  5.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:05<00:07,  7.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:05<00:06,  8.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:05<00:05,  8.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:09,  4.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:07<00:13,  3.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:10<00:28,  1.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:10<00:19,  2.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:10<00:13,  2.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:10<00:12,  3.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:10<00:10,  3.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:11<00:11,  3.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:11<00:11,  2.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:11<00:09,  3.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:11<00:08,  3.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:12<00:07,  4.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:12<00:08,  3.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:12<00:07,  4.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:12<00:06,  4.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:13<00:05,  4.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:13<00:04,  5.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:13<00:06,  3.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:15<00:17,  1.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:15<00:12,  1.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:15<00:09,  2.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:15<00:07,  3.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:15<00:05,  3.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:16<00:04,  4.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:16<00:03,  5.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:16<00:03,  5.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:17<00:07,  2.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:17<00:05,  2.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:17<00:04,  2.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:18<00:03,  3.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:18<00:02,  4.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:18<00:02,  4.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:18<00:01,  5.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:18<00:01,  5.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:18<00:01,  5.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:19<00:01,  3.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:19<00:01,  4.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:19<00:01,  3.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:20<00:01,  3.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:20<00:00,  4.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:20<00:00,  4.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:20<00:00,  4.07it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:20<00:00,  4.79it/s]\n",
            "Scanning stories:   0%|          | 85/739990 [07:06<664:16:36,  3.23s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 18: Found 530 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 34.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:02, 31.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:02, 32.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:02, 32.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:00<00:02, 31.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:00<00:02, 28.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:03, 22.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:03, 21.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:01<00:04, 16.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:01<00:03, 16.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:01<00:04, 14.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:01<00:04, 14.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:02<00:04, 14.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:12,  4.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:04<00:18,  2.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:04<00:13,  3.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:04<00:10,  4.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:05<00:15,  3.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:06<00:11,  4.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:08,  5.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:06<00:07,  5.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:06<00:06,  6.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:06<00:06,  6.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:07<00:05,  7.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:08<00:12,  2.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:10<00:19,  1.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:10<00:15,  2.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:10<00:13,  2.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:10<00:10,  3.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:10<00:08,  3.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:10<00:07,  4.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:11<00:07,  3.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:11<00:06,  4.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:11<00:07,  3.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:05,  4.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:11<00:04,  5.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:12<00:04,  5.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:14<00:19,  1.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:14<00:13,  1.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:14<00:10,  2.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:19<00:34,  1.70s/it]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:19<00:23,  1.23s/it]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:20<00:21,  1.18s/it]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:20<00:14,  1.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:20<00:10,  1.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:20<00:07,  2.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:21<00:06,  2.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:21<00:05,  2.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:21<00:05,  2.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:22<00:03,  2.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:22<00:03,  2.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:22<00:02,  3.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:23<00:02,  2.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:24<00:03,  1.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:24<00:02,  2.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:25<00:04,  1.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:27<00:04,  1.09s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:29<00:03,  1.25s/it]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:29<00:01,  1.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:39<00:03,  3.56s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:39<00:00,  2.54it/s]\n",
            "Scanning stories:   0%|          | 86/739990 [07:46<1132:26:41,  5.51s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 19: Found 559 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:04, 22.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:   7%|▋         | 7/100 [00:00<00:03, 27.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:03, 23.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:04, 18.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:07, 12.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:01<00:06, 12.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:05, 13.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:06, 12.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:07, 10.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:06, 12.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:05, 13.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:03<00:17,  3.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:03<00:15,  4.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:03<00:12,  5.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:04<00:10,  6.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:04<00:09,  6.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:04<00:07,  8.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:08,  6.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:04<00:07,  7.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:05<00:06,  7.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:05<00:05,  9.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:05<00:05,  9.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:05<00:04, 10.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:07<00:12,  3.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:07<00:09,  4.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:07<00:08,  5.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:07<00:06,  6.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:07<00:06,  6.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:08<00:07,  4.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:08<00:11,  3.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:08<00:07,  4.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:07,  4.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:10<00:13,  2.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:10<00:10,  3.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:10<00:08,  3.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:10<00:07,  4.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:10<00:05,  5.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:05,  5.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:10<00:04,  6.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:10<00:03,  6.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:03,  7.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:11<00:04,  5.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:11<00:03,  6.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:13<00:16,  1.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:13<00:11,  1.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:17<00:33,  1.58s/it]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:17<00:22,  1.14s/it]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:17<00:15,  1.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:18<00:12,  1.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:18<00:09,  1.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:18<00:06,  2.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:18<00:06,  2.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:19<00:04,  3.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:19<00:03,  3.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:19<00:04,  2.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:19<00:03,  3.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:20<00:02,  3.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:20<00:02,  3.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:20<00:02,  3.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:25<00:10,  1.50s/it]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:26<00:09,  1.57s/it]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:27<00:07,  1.43s/it]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:28<00:04,  1.05s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:28<00:02,  1.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:30<00:02,  1.13s/it]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:30<00:00,  1.19it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:32<00:00,  3.09it/s]\n",
            "Scanning stories:   0%|          | 87/739990 [08:19<1593:28:04,  7.75s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 20: only 418 tokens (need 1024)\n",
            "Story 20: Found 576 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 34.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:04, 22.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:03, 22.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:05, 16.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:00<00:04, 17.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:04, 19.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:03, 21.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:06, 11.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:05, 13.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:02<00:05, 13.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:04, 14.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:04, 13.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:04, 14.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:03<00:08,  7.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:03<00:06,  8.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:03<00:07,  7.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:03<00:07,  7.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:04<00:07,  7.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:04<00:05,  8.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:04<00:06,  7.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:04<00:07,  6.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:09,  4.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:05<00:07,  5.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:06<00:11,  3.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:06<00:12,  3.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:07<00:09,  4.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:07<00:08,  4.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:07,  5.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:07<00:06,  5.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:07<00:05,  6.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:07<00:05,  5.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:08<00:05,  6.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:08<00:04,  6.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:08<00:04,  7.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:08<00:03,  7.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:08<00:03,  7.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:11<00:24,  1.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:17,  1.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:11<00:13,  1.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:11<00:10,  2.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:16<00:36,  1.58s/it]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:16<00:25,  1.15s/it]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:16<00:17,  1.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:16<00:12,  1.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:16<00:09,  2.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:17<00:08,  2.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:17<00:06,  2.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:17<00:04,  3.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:17<00:03,  3.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:17<00:03,  4.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:21<00:15,  1.17s/it]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:21<00:10,  1.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:21<00:08,  1.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:22<00:05,  1.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:22<00:04,  1.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:22<00:03,  2.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:24<00:05,  1.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:25<00:04,  1.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:25<00:03,  1.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:25<00:01,  2.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:27<00:02,  1.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:27<00:01,  1.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:27<00:00,  1.80it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:28<00:00,  3.56it/s]\n",
            "Scanning stories:   0%|          | 89/739990 [08:48<1851:12:39,  9.01s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 21: only 354 tokens (need 1024)\n",
            "Skipping story 21: only 414 tokens (need 1024)\n",
            "Skipping story 21: only 405 tokens (need 1024)\n",
            "Story 21: Found 537 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:02, 31.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:03, 23.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:05, 15.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:01<00:07, 11.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:11,  7.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:08,  9.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:02<00:06, 10.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:02<00:05, 12.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:05, 13.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:04, 14.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:04, 13.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:04, 14.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:04, 14.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:02<00:03, 15.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:03<00:12,  4.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:04<00:11,  5.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:04<00:08,  6.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:04<00:08,  6.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:05<00:09,  5.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:05<00:09,  5.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:05<00:08,  5.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:05<00:06,  7.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:05<00:08,  5.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:06,  6.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:06<00:13,  3.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:08<00:16,  2.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:08<00:12,  3.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:08<00:13,  2.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:09<00:09,  4.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:09<00:09,  3.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:09,  3.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:08,  4.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:09<00:06,  4.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:10<00:06,  4.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:10<00:05,  5.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:10<00:05,  5.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:06,  4.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:10<00:05,  5.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:10<00:04,  5.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:04,  6.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:11<00:03,  6.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:11<00:03,  7.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:11<00:03,  5.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:11<00:03,  6.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:11<00:03,  6.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:11<00:02,  7.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:12<00:02,  7.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:12<00:03,  4.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:12<00:04,  3.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:13<00:04,  3.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:13<00:03,  4.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:13<00:02,  4.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:13<00:02,  5.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:13<00:02,  5.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:13<00:01,  5.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:14<00:01,  6.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:14<00:01,  6.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:14<00:01,  6.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:15<00:03,  1.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:15<00:02,  2.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:16<00:01,  2.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:18<00:03,  1.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:18<00:01,  1.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:18<00:00,  2.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:18<00:00,  1.89it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:19<00:00,  5.23it/s]\n",
            "Scanning stories:   0%|          | 93/739990 [09:08<1560:27:16,  7.59s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 22: Found 558 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 29.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:   7%|▋         | 7/100 [00:00<00:02, 31.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:02, 31.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:02, 31.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:00<00:02, 29.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:00<00:03, 22.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:00<00:03, 22.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:03, 21.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:01<00:03, 20.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:01<00:03, 18.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:01<00:04, 15.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:01<00:03, 15.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:01<00:03, 15.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:02<00:03, 15.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:02<00:03, 14.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:02<00:03, 14.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:02<00:05,  9.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:02<00:04, 10.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:03<00:04, 10.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:03<00:04,  9.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:03<00:04, 10.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:03<00:03, 10.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:03<00:03, 10.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:04<00:03, 10.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:04<00:03, 10.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:05<00:08,  3.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:05<00:07,  4.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:05<00:06,  4.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:06<00:08,  3.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:06<00:07,  3.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:06<00:06,  4.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:06<00:05,  5.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:06<00:04,  5.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:06<00:04,  6.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:07<00:03,  6.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:07<00:03,  6.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:07<00:03,  7.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:08<00:09,  2.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:08<00:07,  2.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:08<00:05,  3.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:08<00:04,  4.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:08<00:03,  4.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:09<00:03,  4.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:09<00:03,  5.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:09<00:02,  5.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:09<00:02,  6.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:09<00:02,  6.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:09<00:01,  6.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:10<00:01,  6.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:11<00:06,  1.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:11<00:04,  2.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:12<00:03,  2.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:13<00:05,  1.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:14<00:03,  1.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:14<00:02,  2.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:14<00:01,  2.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:14<00:00,  3.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:14<00:00,  3.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:14<00:00,  3.76it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:15<00:00,  6.61it/s]\n",
            "Scanning stories:   0%|          | 94/739990 [09:24<1749:34:48,  8.51s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 23: Found 557 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:02, 32.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:03, 24.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:03, 22.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:00<00:04, 16.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:05, 15.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:07, 10.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:06, 11.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:05, 12.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:05, 14.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:04, 15.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:02<00:05, 11.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:07,  8.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:07,  8.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:06, 10.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:02<00:05, 11.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:03<00:04, 11.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:03<00:04, 12.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:03<00:04, 12.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:03<00:05, 10.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:04<00:06,  7.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:05<00:16,  2.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:12,  3.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:14,  3.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:06<00:10,  4.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:06<00:09,  4.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:07<00:07,  5.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:07<00:06,  6.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:05,  7.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:07<00:04,  7.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:07<00:04,  8.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:08<00:07,  4.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:08<00:06,  4.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:10<00:22,  1.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:12<00:24,  1.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:12<00:18,  1.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:12<00:13,  1.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:12<00:10,  2.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:12<00:08,  2.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:12<00:06,  3.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:13<00:05,  4.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:13<00:04,  4.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:13<00:03,  5.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:15<00:13,  1.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:15<00:09,  1.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:15<00:07,  2.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:15<00:05,  3.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:15<00:04,  3.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:15<00:03,  4.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:15<00:02,  5.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:16<00:03,  3.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:16<00:03,  3.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:16<00:02,  4.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:16<00:02,  4.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:17<00:03,  2.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:17<00:02,  2.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:18<00:02,  2.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:18<00:01,  3.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:18<00:01,  3.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:18<00:00,  4.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:19<00:00,  4.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:19<00:00,  5.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:19<00:00,  5.36it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:19<00:00,  5.13it/s]\n",
            "Scanning stories:   0%|          | 95/739990 [09:44<2079:13:48, 10.12s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 24: Found 537 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 29.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:04, 21.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:06, 15.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:04, 17.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:01<00:06, 12.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:01<00:08, 10.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:06, 12.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:05, 14.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:05, 13.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:05, 12.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:07,  9.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:06, 10.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:05, 11.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:03<00:11,  5.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:09,  6.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:03<00:07,  7.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:04<00:06,  9.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:04<00:07,  6.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:04<00:06,  8.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:04<00:06,  7.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:05<00:06,  7.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:05<00:05,  8.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:04,  9.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:07<00:16,  2.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:07<00:14,  3.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:07<00:10,  4.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:08<00:09,  4.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:09<00:14,  2.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:10,  3.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:09,  3.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:10<00:09,  3.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:12<00:25,  1.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:13<00:20,  1.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:13<00:15,  1.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:13<00:11,  2.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:13<00:09,  3.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:13<00:07,  3.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:13<00:05,  4.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:13<00:04,  5.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:13<00:05,  4.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:14<00:04,  5.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:14<00:04,  4.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:14<00:03,  5.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:14<00:04,  4.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:14<00:03,  5.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:15<00:03,  5.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:15<00:02,  6.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:15<00:03,  5.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:16<00:05,  2.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:16<00:04,  3.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:19<00:14,  1.13s/it]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:19<00:09,  1.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:21<00:13,  1.18s/it]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:21<00:08,  1.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:21<00:05,  1.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:22<00:04,  1.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:22<00:02,  2.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:22<00:01,  3.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:22<00:01,  3.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:22<00:00,  4.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:29<00:06,  2.26s/it]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:29<00:03,  1.63s/it]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:29<00:01,  1.19s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:30<00:00,  3.32it/s]\n",
            "Scanning stories:   0%|          | 96/739990 [10:15<2785:23:30, 13.55s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 25: only 215 tokens (need 1024)\n",
            "Story 25: Found 561 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:04, 23.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:03, 25.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:05, 17.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:04, 19.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:04, 20.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:01<00:07, 10.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:07, 11.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:05, 13.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:04, 15.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:04, 16.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:04, 17.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:02<00:04, 15.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:05, 12.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:04, 13.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:03<00:10,  5.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:04<00:15,  3.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:04<00:12,  4.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:04<00:09,  5.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:05<00:17,  3.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:05<00:12,  4.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:05<00:09,  5.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:06<00:11,  4.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:07<00:16,  3.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:07<00:11,  4.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:07<00:12,  3.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:08<00:10,  4.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:08<00:07,  5.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:08<00:06,  6.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:08<00:04,  7.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:04,  8.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:06,  5.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:09<00:07,  4.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:10<00:06,  4.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:10<00:05,  5.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:10<00:05,  5.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:04,  6.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:10<00:04,  6.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:10<00:03,  7.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:10<00:03,  7.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:10<00:03,  7.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:10<00:03,  7.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:11<00:02,  7.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:11<00:02,  7.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:11<00:02,  8.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:14<00:21,  1.09s/it]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:14<00:15,  1.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:15<00:12,  1.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:15<00:08,  1.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:15<00:06,  2.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:15<00:04,  3.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:16<00:06,  2.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:16<00:04,  2.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:16<00:03,  3.36it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:16<00:02,  4.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:17<00:03,  2.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:17<00:02,  3.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:18<00:03,  2.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:18<00:02,  2.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:20<00:05,  1.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:21<00:03,  1.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:21<00:02,  1.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:21<00:01,  2.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:21<00:00,  2.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:21<00:00,  3.23it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:21<00:00,  4.55it/s]\n",
            "Scanning stories:   0%|          | 98/739990 [10:38<2643:18:30, 12.86s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 26: only 355 tokens (need 1024)\n",
            "Skipping story 26: only 728 tokens (need 1024)\n",
            "Skipping story 26: only 540 tokens (need 1024)\n",
            "Story 26: Found 512 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:03, 31.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:03, 29.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:03, 26.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:05, 15.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:05, 14.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:04, 16.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:05, 13.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:04, 15.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:06, 10.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:12,  5.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:11,  6.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:03<00:08,  7.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:03<00:07,  9.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:03<00:05, 10.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:04<00:14,  4.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:10,  5.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:04<00:09,  5.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:05<00:08,  6.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:05<00:06,  7.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:06<00:12,  4.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:06<00:09,  5.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:06<00:08,  5.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:06<00:06,  6.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:06<00:05,  7.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:06<00:04,  8.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:07<00:06,  6.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:08<00:12,  3.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:08<00:08,  4.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:08<00:07,  4.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:08<00:06,  5.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:05,  5.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:09<00:05,  6.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:09<00:09,  3.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:11<00:19,  1.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:11<00:14,  2.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:11<00:11,  2.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:11<00:08,  3.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:11<00:06,  3.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:13<00:20,  1.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:14<00:15,  1.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:14<00:11,  2.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:14<00:08,  2.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:14<00:06,  3.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:14<00:05,  4.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:15<00:05,  3.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:15<00:04,  4.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:15<00:03,  4.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:15<00:03,  5.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:15<00:02,  5.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:15<00:02,  5.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:15<00:02,  5.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:16<00:02,  5.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:17<00:06,  1.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:17<00:04,  2.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:17<00:03,  3.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:17<00:02,  3.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:18<00:01,  4.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:18<00:01,  4.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:18<00:01,  5.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:18<00:01,  3.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:18<00:00,  4.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:19<00:00,  4.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:19<00:00,  4.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:19<00:00,  4.24it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:19<00:00,  5.06it/s]\n",
            "Scanning stories:   0%|          | 102/739990 [10:59<1886:10:14,  9.18s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 27: only 495 tokens (need 1024)\n",
            "Story 27: Found 540 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:04, 22.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:04, 18.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:04, 20.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:00<00:05, 16.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:04, 16.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:06, 12.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:08,  9.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:07,  9.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:02<00:07,  9.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:06, 10.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:07,  9.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:08,  8.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:07,  8.36it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:03<00:06, 10.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:03<00:05, 11.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:03<00:06,  9.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:03<00:06,  8.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:04<00:07,  7.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:04<00:06,  8.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:04<00:05,  9.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:04<00:04, 10.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:04<00:04,  9.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:07,  6.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:06<00:17,  2.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:07<00:14,  3.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:07<00:12,  3.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:08<00:16,  2.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:09<00:20,  1.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:09<00:17,  2.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:09<00:11,  3.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:09<00:07,  4.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:11<00:15,  2.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:11<00:15,  2.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:11<00:12,  2.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:11<00:10,  3.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:11<00:08,  3.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:12<00:06,  4.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:12<00:05,  5.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:12<00:05,  4.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:12<00:04,  5.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:12<00:04,  6.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:12<00:03,  6.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:12<00:03,  6.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:15<00:16,  1.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:15<00:11,  1.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:15<00:08,  2.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:15<00:06,  2.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:15<00:05,  3.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:15<00:04,  4.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:15<00:03,  4.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:18<00:15,  1.04s/it]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:19<00:10,  1.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:19<00:07,  1.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:19<00:05,  2.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:19<00:04,  2.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:19<00:03,  3.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:19<00:02,  3.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:19<00:01,  4.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:20<00:01,  4.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:20<00:01,  5.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:23<00:05,  1.20s/it]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:25<00:05,  1.37s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:25<00:03,  1.01s/it]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:25<00:01,  1.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:26<00:00,  1.60it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:28<00:00,  3.55it/s]\n",
            "Scanning stories:   0%|          | 104/739990 [11:28<2163:26:33, 10.53s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 28: only 263 tokens (need 1024)\n",
            "Skipping story 28: only 678 tokens (need 1024)\n",
            "Story 28: Found 526 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 25.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:05, 18.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:04, 19.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:06, 14.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  15%|█▌        | 15/100 [00:00<00:05, 16.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:01<00:05, 14.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:06, 11.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:01<00:09,  8.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:06, 10.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:02<00:05, 13.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:07,  9.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:06, 11.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:05, 11.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:05, 11.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:03<00:08,  7.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:08,  7.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:03<00:08,  7.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:14,  4.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:04<00:13,  4.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:05<00:24,  2.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:05<00:15,  3.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:06<00:14,  3.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:06<00:13,  3.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:06<00:13,  3.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:06<00:13,  3.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:07<00:11,  4.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:08<00:24,  1.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:08<00:15,  2.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:09<00:11,  3.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:09<00:08,  5.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:09<00:06,  6.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:09<00:05,  7.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:09<00:04,  7.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:09<00:05,  6.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:10<00:05,  6.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:10<00:04,  7.36it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:10<00:04,  7.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:10<00:04,  7.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:10<00:03,  8.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:10<00:03,  8.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:04,  6.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:11<00:12,  2.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:12<00:09,  2.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:12<00:07,  3.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:12<00:05,  4.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:12<00:04,  5.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:12<00:04,  5.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:16<00:28,  1.29s/it]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:16<00:19,  1.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:16<00:13,  1.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:16<00:10,  1.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:17<00:08,  2.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:17<00:07,  2.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:17<00:06,  2.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:20<00:15,  1.03s/it]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:22<00:20,  1.50s/it]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:23<00:14,  1.10s/it]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:23<00:09,  1.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:23<00:06,  1.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:23<00:04,  2.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:23<00:03,  2.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:24<00:02,  2.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:24<00:02,  3.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:24<00:01,  4.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:24<00:01,  4.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:24<00:00,  5.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:24<00:00,  4.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:25<00:00,  3.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:30<00:01,  1.63s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:30<00:00,  3.26it/s]\n",
            "Scanning stories:   0%|          | 107/739990 [11:59<2162:13:12, 10.52s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 29: Found 521 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:03, 31.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:03, 23.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:04, 21.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:04, 18.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:06, 13.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:01<00:07, 10.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:07, 10.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:07, 10.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:06, 11.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:05, 14.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:08,  8.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:07,  8.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:06, 10.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:05, 11.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:04, 12.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:05, 11.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:03<00:05, 10.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:05, 11.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:03<00:05,  9.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:03<00:05, 10.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:04<00:05,  9.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:04<00:05,  9.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:05<00:10,  4.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:05<00:08,  5.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:05<00:06,  6.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:06<00:06,  6.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:08<00:17,  2.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:08<00:12,  2.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:08<00:11,  3.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:08<00:09,  3.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:09<00:08,  4.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:09<00:06,  4.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:09<00:05,  5.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:09<00:05,  6.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:09<00:05,  5.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:14,  1.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:11<00:11,  2.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:11<00:08,  3.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:12<00:17,  1.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:12<00:12,  1.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:12<00:09,  2.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:13<00:07,  3.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:13<00:05,  3.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:13<00:04,  4.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:13<00:03,  5.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:13<00:03,  5.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:14<00:04,  4.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:14<00:06,  2.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:14<00:05,  2.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:15<00:04,  3.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:15<00:03,  4.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:16<00:07,  1.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:16<00:05,  2.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:16<00:03,  2.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:17<00:02,  3.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:17<00:02,  3.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:17<00:01,  4.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:17<00:01,  4.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:17<00:01,  5.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:18<00:02,  2.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:18<00:01,  2.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:19<00:00,  3.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:19<00:00,  3.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:19<00:00,  3.76it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:19<00:00,  5.07it/s]\n",
            "Scanning stories:   0%|          | 108/739990 [12:20<2456:30:16, 11.95s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 30: only 183 tokens (need 1024)\n",
            "Story 30: Found 586 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 29.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:   7%|▋         | 7/100 [00:00<00:03, 26.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:04, 20.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:04, 20.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:03, 23.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:00<00:03, 24.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:00<00:03, 22.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:03, 22.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  28%|██▊       | 28/100 [00:01<00:04, 15.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:04, 16.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:04, 13.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:02<00:06, 10.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:07,  8.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:10,  6.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:03<00:08,  7.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:03<00:06,  8.55it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:03<00:06,  8.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:03<00:06,  8.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:03<00:06,  8.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:04<00:05,  9.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:04<00:06,  8.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:04<00:05,  9.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:04<00:05,  8.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:15,  3.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:05<00:10,  4.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:06<00:15,  2.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:06<00:10,  3.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:07<00:12,  3.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:07<00:08,  4.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:07<00:09,  3.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:07,  4.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:08<00:12,  2.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:08<00:09,  3.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:11<00:30,  1.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:11<00:24,  1.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:11<00:17,  1.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:12<00:16,  1.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:12<00:14,  2.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:12<00:11,  2.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:13<00:09,  2.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:13<00:07,  3.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:13<00:07,  3.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:14<00:13,  1.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:14<00:09,  2.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:15<00:07,  2.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:15<00:05,  3.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:16<00:08,  2.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:16<00:06,  2.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:16<00:05,  3.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:23<00:37,  2.23s/it]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:23<00:25,  1.60s/it]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:23<00:17,  1.16s/it]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:23<00:11,  1.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:24<00:12,  1.03it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:25<00:12,  1.05s/it]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:26<00:08,  1.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:26<00:05,  1.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:27<00:06,  1.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:27<00:04,  1.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:35<00:20,  2.90s/it]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:36<00:13,  2.21s/it]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:44<00:20,  4.11s/it]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:45<00:11,  2.93s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:45<00:06,  2.10s/it]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:45<00:03,  1.56s/it]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:45<00:01,  1.14s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:47<00:00,  2.13it/s]\n",
            "Scanning stories:   0%|          | 110/739990 [13:07<3169:53:12, 15.42s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 31: only 919 tokens (need 1024)\n",
            "Skipping story 31: only 785 tokens (need 1024)\n",
            "Skipping story 31: only 288 tokens (need 1024)\n",
            "Skipping story 31: only 209 tokens (need 1024)\n",
            "Story 31: Found 571 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:04, 23.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:04, 20.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  10%|█         | 10/100 [00:00<00:03, 25.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:04, 20.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:03, 22.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:00<00:03, 22.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:05, 15.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:02<00:10,  7.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:02<00:11,  6.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:09,  7.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:08,  8.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:07,  9.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:03<00:07,  8.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:03<00:06, 10.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:05, 11.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:03<00:06,  9.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:05,  9.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:04<00:12,  4.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:05<00:13,  4.11it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:05<00:09,  5.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:05<00:07,  6.57it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:05<00:06,  7.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:05,  7.88it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:06<00:05,  8.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:06<00:04,  9.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:06<00:05,  7.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:06<00:04,  8.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:06<00:04,  8.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:04,  8.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:07<00:03,  8.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:07<00:03,  8.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:07<00:05,  5.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:07<00:04,  6.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:07<00:04,  6.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:08<00:06,  4.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:08<00:05,  5.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:08<00:04,  6.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:08<00:04,  6.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:08<00:03,  7.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:08<00:03,  7.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:08<00:03,  7.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:09<00:02,  7.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:09<00:02,  7.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:09<00:02,  7.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:09<00:02,  7.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:10<00:09,  2.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:10<00:06,  2.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:11<00:05,  3.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:11<00:04,  3.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:11<00:03,  4.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:11<00:03,  4.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:11<00:02,  4.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:11<00:02,  5.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:11<00:01,  5.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:12<00:01,  5.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:12<00:02,  3.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:12<00:02,  3.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:13<00:01,  4.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:13<00:01,  3.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:13<00:01,  4.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:14<00:01,  2.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:14<00:01,  2.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:18<00:02,  1.43s/it]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:18<00:01,  1.05s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:22<00:00,  4.42it/s]\n",
            "Scanning stories:   0%|          | 115/739990 [13:32<2061:05:27, 10.03s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 32: only 714 tokens (need 1024)\n",
            "Skipping story 32: only 601 tokens (need 1024)\n",
            "Story 32: Found 518 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 29.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:04, 21.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:04, 21.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:03, 23.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:04, 17.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:00<00:04, 16.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:01<00:05, 15.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:04, 16.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:07, 10.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:02<00:14,  5.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:11,  6.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:09,  7.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:03<00:10,  6.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:03<00:08,  7.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:03<00:06,  9.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:07,  7.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:04<00:12,  4.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:04<00:11,  4.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:05<00:11,  4.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:05<00:10,  5.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:06<00:21,  2.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:06<00:20,  2.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:07<00:20,  2.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:07<00:24,  2.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:07<00:14,  3.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:08<00:10,  4.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:08<00:09,  4.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:08<00:06,  6.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:08<00:09,  4.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:08<00:06,  6.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:09<00:11,  3.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:10<00:10,  3.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:10<00:07,  4.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:10<00:08,  4.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:11<00:07,  4.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:11<00:07,  4.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:12<00:11,  2.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:12<00:10,  3.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:12<00:08,  3.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:13<00:10,  2.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:13<00:08,  3.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:13<00:06,  4.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:13<00:05,  4.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:14<00:12,  1.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:14<00:09,  2.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:15<00:07,  3.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:15<00:05,  3.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:16<00:15,  1.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:17<00:10,  1.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:17<00:09,  2.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:17<00:07,  2.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:17<00:05,  3.19it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:17<00:04,  3.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:17<00:03,  4.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:18<00:02,  5.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:18<00:02,  5.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:19<00:06,  1.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:20<00:05,  2.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:20<00:05,  1.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:21<00:06,  1.43it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:23<00:08,  1.01s/it]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:25<00:08,  1.23s/it]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:25<00:06,  1.05s/it]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:26<00:03,  1.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:27<00:03,  1.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:27<00:02,  1.35it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:27<00:01,  1.76it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:30<00:01,  1.08s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:30<00:00,  3.30it/s]\n",
            "Scanning stories:   0%|          | 118/739990 [14:03<2081:31:46, 10.13s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 33: only 402 tokens (need 1024)\n",
            "Story 33: Found 541 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:04, 23.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:03, 24.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:03, 25.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:04, 17.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:04, 17.87it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:00<00:03, 21.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:03, 22.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:03, 20.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:01<00:03, 20.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:05, 12.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:01<00:04, 13.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:02<00:04, 14.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:04, 14.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:03<00:12,  4.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:03<00:09,  5.97it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:03<00:09,  5.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:04<00:07,  6.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:04<00:06,  7.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:04<00:06,  7.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:04<00:06,  7.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:05<00:07,  6.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:05<00:06,  7.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:05<00:05,  8.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:05<00:07,  5.64it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:05<00:05,  6.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:06<00:04,  7.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:06<00:04,  8.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:06<00:04,  8.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:06<00:04,  8.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:06<00:03,  8.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:06<00:03,  8.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:06<00:04,  6.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:07<00:04,  7.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:07<00:06,  4.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:07<00:05,  5.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:07<00:04,  6.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:07<00:04,  6.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:07<00:03,  6.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:08<00:05,  4.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:08<00:04,  5.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:08<00:03,  5.95it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:10<00:15,  1.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:10<00:11,  1.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:10<00:08,  2.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:10<00:06,  3.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:11<00:04,  3.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:12<00:12,  1.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:13<00:09,  1.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:13<00:08,  1.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:13<00:06,  2.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:14<00:05,  2.33it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:14<00:04,  2.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:14<00:04,  2.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:15<00:03,  3.27it/s]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:15<00:04,  1.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:16<00:03,  2.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:16<00:02,  3.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:16<00:01,  3.61it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:17<00:02,  2.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:17<00:01,  2.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:17<00:00,  3.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:17<00:00,  3.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:22<00:01,  1.63s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:23<00:00,  4.19it/s]\n",
            "Scanning stories:   0%|          | 120/739990 [14:27<2182:10:01, 10.62s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 34: Found 514 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:03, 31.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:04, 21.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:03, 22.30it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:03, 21.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  18%|█▊        | 18/100 [00:00<00:03, 24.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  21%|██        | 21/100 [00:00<00:03, 20.65it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:03, 21.69it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:03, 22.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  30%|███       | 30/100 [00:01<00:03, 17.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:04, 16.22it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:01<00:03, 16.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:01<00:03, 17.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:01<00:03, 17.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:02<00:04, 12.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:02<00:04, 13.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:02<00:04, 13.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:03<00:10,  5.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:03<00:08,  6.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:03<00:06,  7.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:03<00:05,  8.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:04<00:07,  6.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:04<00:06,  6.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:04<00:05,  7.73it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:04<00:04,  8.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:05<00:04,  9.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:05<00:03,  9.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:05<00:05,  6.47it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:06<00:04,  6.84it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:06<00:04,  7.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:06<00:04,  7.52it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:06<00:05,  5.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:08<00:16,  1.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:08<00:12,  2.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:08<00:11,  2.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:08<00:08,  3.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:08<00:06,  3.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:09<00:05,  4.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:09<00:04,  5.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:10<00:09,  2.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:10<00:07,  2.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:10<00:06,  3.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:10<00:04,  3.99it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:10<00:03,  4.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:10<00:03,  5.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:11<00:05,  3.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:16<00:25,  1.71s/it]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:16<00:18,  1.29s/it]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:16<00:12,  1.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:17<00:08,  1.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:20<00:14,  1.36s/it]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:20<00:10,  1.04s/it]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:20<00:06,  1.29it/s]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:20<00:04,  1.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:20<00:03,  2.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:25<00:10,  1.81s/it]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:25<00:06,  1.32s/it]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:27<00:05,  1.26s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:27<00:02,  1.02it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:28<00:02,  1.02s/it]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:28<00:00,  1.31it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:28<00:00,  3.47it/s]\n",
            "Scanning stories:   0%|          | 121/739990 [14:57<2722:11:23, 13.25s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 35: only 213 tokens (need 1024)\n",
            "Story 35: Found 564 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:03, 29.68it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:03, 27.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:04, 22.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  12%|█▏        | 12/100 [00:00<00:03, 24.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:03, 27.26it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:01<00:06, 12.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:05, 13.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  24%|██▍       | 24/100 [00:01<00:07, 10.48it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:05, 12.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:01<00:05, 13.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:05, 13.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:08,  8.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:07,  8.51it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:03<00:07,  8.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:03<00:06,  8.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:03<00:07,  7.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:06,  8.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:03<00:05,  9.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:03<00:04, 10.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:04<00:05, 10.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:04<00:04, 10.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:05<00:11,  4.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:06<00:19,  2.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:07<00:13,  3.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  58%|█████▊    | 58/100 [00:07<00:10,  4.14it/s]\u001b[A\n",
            "Checking tokens for context lengths:  60%|██████    | 60/100 [00:07<00:08,  4.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  62%|██████▏   | 62/100 [00:07<00:06,  5.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:07<00:06,  6.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:05,  6.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:08<00:04,  7.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:08<00:04,  7.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:08<00:04,  7.74it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:09<00:10,  2.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:09<00:10,  2.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:09<00:08,  3.62it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:10<00:16,  1.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:11<00:12,  2.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:11<00:09,  2.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:11<00:07,  3.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:11<00:05,  4.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:11<00:04,  4.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:11<00:04,  5.45it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:12<00:06,  3.60it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:12<00:05,  3.67it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:12<00:04,  4.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:12<00:03,  5.00it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:12<00:03,  5.56it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:13<00:02,  6.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:14<00:08,  1.93it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:14<00:07,  1.91it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:15<00:05,  2.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:15<00:04,  3.07it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:15<00:03,  3.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:15<00:03,  3.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:21<00:20,  2.02s/it]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:22<00:16,  1.80s/it]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:23<00:10,  1.31s/it]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:23<00:06,  1.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:23<00:04,  1.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:23<00:03,  1.53it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:24<00:02,  1.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:24<00:01,  2.49it/s]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:24<00:00,  3.01it/s]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:24<00:00,  3.56it/s]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:24<00:00,  4.04it/s]\n",
            "Scanning stories:   0%|          | 123/739990 [15:22<2694:15:17, 13.11s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Skipping story 36: only 582 tokens (need 1024)\n",
            "Skipping story 36: only 201 tokens (need 1024)\n",
            "Skipping story 36: only 482 tokens (need 1024)\n",
            "Story 36: Found 557 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   4%|▍         | 4/100 [00:00<00:02, 32.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:   8%|▊         | 8/100 [00:00<00:04, 22.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  11%|█         | 11/100 [00:00<00:04, 19.63it/s]\u001b[A\n",
            "Checking tokens for context lengths:  14%|█▍        | 14/100 [00:00<00:03, 21.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  17%|█▋        | 17/100 [00:00<00:03, 23.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  20%|██        | 20/100 [00:00<00:03, 22.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  23%|██▎       | 23/100 [00:01<00:04, 17.37it/s]\u001b[A\n",
            "Checking tokens for context lengths:  26%|██▌       | 26/100 [00:01<00:03, 18.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:01<00:03, 19.28it/s]\u001b[A\n",
            "Checking tokens for context lengths:  32%|███▏      | 32/100 [00:01<00:04, 16.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  34%|███▍      | 34/100 [00:01<00:03, 16.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  36%|███▌      | 36/100 [00:01<00:04, 15.42it/s]\u001b[A\n",
            "Checking tokens for context lengths:  38%|███▊      | 38/100 [00:02<00:03, 15.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  40%|████      | 40/100 [00:02<00:08,  6.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  42%|████▏     | 42/100 [00:04<00:19,  2.94it/s]\u001b[A\n",
            "Checking tokens for context lengths:  44%|████▍     | 44/100 [00:04<00:14,  3.79it/s]\u001b[A\n",
            "Checking tokens for context lengths:  46%|████▌     | 46/100 [00:04<00:11,  4.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  48%|████▊     | 48/100 [00:04<00:08,  5.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  50%|█████     | 50/100 [00:05<00:07,  6.96it/s]\u001b[A\n",
            "Checking tokens for context lengths:  52%|█████▏    | 52/100 [00:05<00:07,  6.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  54%|█████▍    | 54/100 [00:05<00:07,  6.50it/s]\u001b[A\n",
            "Checking tokens for context lengths:  56%|█████▌    | 56/100 [00:05<00:06,  6.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:06<00:06,  6.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  59%|█████▉    | 59/100 [00:06<00:05,  7.77it/s]\u001b[A\n",
            "Checking tokens for context lengths:  61%|██████    | 61/100 [00:06<00:05,  6.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  63%|██████▎   | 63/100 [00:06<00:04,  7.59it/s]\u001b[A\n",
            "Checking tokens for context lengths:  64%|██████▍   | 64/100 [00:07<00:04,  7.85it/s]\u001b[A\n",
            "Checking tokens for context lengths:  65%|██████▌   | 65/100 [00:07<00:09,  3.83it/s]\u001b[A\n",
            "Checking tokens for context lengths:  66%|██████▌   | 66/100 [00:07<00:07,  4.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  67%|██████▋   | 67/100 [00:08<00:07,  4.41it/s]\u001b[A\n",
            "Checking tokens for context lengths:  68%|██████▊   | 68/100 [00:08<00:06,  5.04it/s]\u001b[A\n",
            "Checking tokens for context lengths:  69%|██████▉   | 69/100 [00:08<00:06,  4.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  70%|███████   | 70/100 [00:08<00:07,  3.98it/s]\u001b[A\n",
            "Checking tokens for context lengths:  71%|███████   | 71/100 [00:09<00:09,  3.13it/s]\u001b[A\n",
            "Checking tokens for context lengths:  72%|███████▏  | 72/100 [00:09<00:10,  2.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  73%|███████▎  | 73/100 [00:09<00:07,  3.38it/s]\u001b[A\n",
            "Checking tokens for context lengths:  74%|███████▍  | 74/100 [00:10<00:06,  4.08it/s]\u001b[A\n",
            "Checking tokens for context lengths:  75%|███████▌  | 75/100 [00:10<00:06,  4.09it/s]\u001b[A\n",
            "Checking tokens for context lengths:  76%|███████▌  | 76/100 [00:10<00:04,  4.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  77%|███████▋  | 77/100 [00:12<00:18,  1.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  78%|███████▊  | 78/100 [00:13<00:20,  1.06it/s]\u001b[A\n",
            "Checking tokens for context lengths:  79%|███████▉  | 79/100 [00:14<00:14,  1.44it/s]\u001b[A\n",
            "Checking tokens for context lengths:  80%|████████  | 80/100 [00:14<00:10,  1.90it/s]\u001b[A\n",
            "Checking tokens for context lengths:  81%|████████  | 81/100 [00:14<00:07,  2.46it/s]\u001b[A\n",
            "Checking tokens for context lengths:  82%|████████▏ | 82/100 [00:14<00:05,  3.10it/s]\u001b[A\n",
            "Checking tokens for context lengths:  83%|████████▎ | 83/100 [00:15<00:10,  1.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  84%|████████▍ | 84/100 [00:15<00:07,  2.21it/s]\u001b[A\n",
            "Checking tokens for context lengths:  85%|████████▌ | 85/100 [00:15<00:05,  2.80it/s]\u001b[A\n",
            "Checking tokens for context lengths:  86%|████████▌ | 86/100 [00:16<00:07,  1.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  87%|████████▋ | 87/100 [00:17<00:05,  2.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  88%|████████▊ | 88/100 [00:17<00:05,  2.18it/s]\u001b[A\n",
            "Checking tokens for context lengths:  89%|████████▉ | 89/100 [00:17<00:04,  2.72it/s]\u001b[A\n",
            "Checking tokens for context lengths:  90%|█████████ | 90/100 [00:20<00:11,  1.13s/it]\u001b[A\n",
            "Checking tokens for context lengths:  91%|█████████ | 91/100 [00:21<00:09,  1.03s/it]\u001b[A\n",
            "Checking tokens for context lengths:  92%|█████████▏| 92/100 [00:22<00:07,  1.05it/s]\u001b[A\n",
            "Checking tokens for context lengths:  93%|█████████▎| 93/100 [00:22<00:05,  1.39it/s]\u001b[A\n",
            "Checking tokens for context lengths:  94%|█████████▍| 94/100 [00:22<00:03,  1.82it/s]\u001b[A\n",
            "Checking tokens for context lengths:  95%|█████████▌| 95/100 [00:22<00:02,  2.31it/s]\u001b[A\n",
            "Checking tokens for context lengths:  96%|█████████▌| 96/100 [00:28<00:08,  2.11s/it]\u001b[A\n",
            "Checking tokens for context lengths:  97%|█████████▋| 97/100 [00:28<00:04,  1.53s/it]\u001b[A\n",
            "Checking tokens for context lengths:  98%|█████████▊| 98/100 [00:31<00:03,  1.79s/it]\u001b[A\n",
            "Checking tokens for context lengths:  99%|█████████▉| 99/100 [00:31<00:01,  1.30s/it]\u001b[A\n",
            "Checking tokens for context lengths: 100%|██████████| 100/100 [00:36<00:00,  2.76it/s]\n",
            "Scanning stories:   0%|          | 127/739990 [15:59<2339:54:44, 11.39s/it]"
          ]
        },
        {
          "output_type": "stream",
          "name": "stdout",
          "text": [
            "Story 37: Found 550 confident tokens, analyzing 100 samples\n"
          ]
        },
        {
          "output_type": "stream",
          "name": "stderr",
          "text": [
            "\n",
            "Checking tokens for context lengths:   0%|          | 0/100 [00:00<?, ?it/s]\u001b[A\n",
            "Checking tokens for context lengths:   3%|▎         | 3/100 [00:00<00:04, 23.75it/s]\u001b[A\n",
            "Checking tokens for context lengths:   6%|▌         | 6/100 [00:00<00:04, 20.25it/s]\u001b[A\n",
            "Checking tokens for context lengths:   9%|▉         | 9/100 [00:00<00:04, 21.32it/s]\u001b[A\n",
            "Checking tokens for context lengths:  13%|█▎        | 13/100 [00:00<00:03, 25.16it/s]\u001b[A\n",
            "Checking tokens for context lengths:  16%|█▌        | 16/100 [00:00<00:03, 26.24it/s]\u001b[A\n",
            "Checking tokens for context lengths:  19%|█▉        | 19/100 [00:00<00:03, 26.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  22%|██▏       | 22/100 [00:01<00:04, 18.15it/s]\u001b[A\n",
            "Checking tokens for context lengths:  25%|██▌       | 25/100 [00:01<00:06, 11.23it/s]\u001b[A\n",
            "Checking tokens for context lengths:  27%|██▋       | 27/100 [00:01<00:06, 11.40it/s]\u001b[A\n",
            "Checking tokens for context lengths:  29%|██▉       | 29/100 [00:02<00:10,  6.81it/s]\u001b[A\n",
            "Checking tokens for context lengths:  31%|███       | 31/100 [00:02<00:08,  8.20it/s]\u001b[A\n",
            "Checking tokens for context lengths:  33%|███▎      | 33/100 [00:02<00:06,  9.66it/s]\u001b[A\n",
            "Checking tokens for context lengths:  35%|███▌      | 35/100 [00:02<00:05, 11.17it/s]\u001b[A\n",
            "Checking tokens for context lengths:  37%|███▋      | 37/100 [00:02<00:05, 12.58it/s]\u001b[A\n",
            "Checking tokens for context lengths:  39%|███▉      | 39/100 [00:02<00:05, 11.70it/s]\u001b[A\n",
            "Checking tokens for context lengths:  41%|████      | 41/100 [00:03<00:05, 11.12it/s]\u001b[A\n",
            "Checking tokens for context lengths:  43%|████▎     | 43/100 [00:03<00:04, 11.89it/s]\u001b[A\n",
            "Checking tokens for context lengths:  45%|████▌     | 45/100 [00:03<00:04, 12.34it/s]\u001b[A\n",
            "Checking tokens for context lengths:  47%|████▋     | 47/100 [00:03<00:04, 12.54it/s]\u001b[A\n",
            "Checking tokens for context lengths:  49%|████▉     | 49/100 [00:03<00:03, 12.78it/s]\u001b[A\n",
            "Checking tokens for context lengths:  51%|█████     | 51/100 [00:04<00:06,  7.92it/s]\u001b[A\n",
            "Checking tokens for context lengths:  53%|█████▎    | 53/100 [00:04<00:05,  8.86it/s]\u001b[A\n",
            "Checking tokens for context lengths:  55%|█████▌    | 55/100 [00:05<00:09,  4.71it/s]\u001b[A\n",
            "Checking tokens for context lengths:  57%|█████▋    | 57/100 [00:06<00:14,  3.04it/s]\u001b[A"
          ]
        }
      ]
    }
  ]
}