chore: refresh tokens and sync runtime state snapshots
This commit is contained in:
@@ -1 +1 @@
|
||||
{"agent_mode_auto_approval": true, "annotations_enabled": true, "azure_only": false, "blackbird_clientside_indexing": false, "chat_enabled": true, "chat_jetbrains_enabled": true, "code_quote_enabled": true, "code_review_enabled": true, "codesearch": true, "copilotignore_enabled": false, "endpoints": {"api": "https://api.individual.githubcopilot.com", "origin-tracker": "https://origin-tracker.individual.githubcopilot.com", "proxy": "https://proxy.individual.githubcopilot.com", "telemetry": "https://telemetry.individual.githubcopilot.com"}, "expires_at": 1773960905, "individual": true, "limited_user_quotas": null, "limited_user_reset_date": null, "prompt_8k": true, "public_suggestions": "disabled", "refresh_in": 1500, "sku": "monthly_subscriber_quota", "snippy_load_test_enabled": false, "telemetry": "disabled", "token": "tid=ded1d75350f66adcb3d0ab36e8e78c47;exp=1773960905;sku=monthly_subscriber_quota;proxy-ep=proxy.individual.githubcopilot.com;st=dotcom;chat=1;cit=1;malfil=1;editor_preview_features=1;agent_mode=1;agent_mode_auto_approval=1;mcp=1;ccr=1;8kp=1;ip=24.143.97.87;asn=AS11404:86857143350a691faff7b3c0e874df67a55579a5f9ef300a285fd27ec8f6bab0", "tracking_id": "ded1d75350f66adcb3d0ab36e8e78c47", "vsc_electron_fetcher_v2": false, "xcode": true, "xcode_chat": false}
|
||||
{"agent_mode_auto_approval": true, "annotations_enabled": true, "azure_only": false, "blackbird_clientside_indexing": false, "chat_enabled": true, "chat_jetbrains_enabled": true, "code_quote_enabled": true, "code_review_enabled": true, "codesearch": true, "copilotignore_enabled": false, "endpoints": {"api": "https://api.individual.githubcopilot.com", "origin-tracker": "https://origin-tracker.individual.githubcopilot.com", "proxy": "https://proxy.individual.githubcopilot.com", "telemetry": "https://telemetry.individual.githubcopilot.com"}, "expires_at": 1774026494, "individual": true, "limited_user_quotas": null, "limited_user_reset_date": null, "prompt_8k": true, "public_suggestions": "disabled", "refresh_in": 1500, "sku": "monthly_subscriber_quota", "snippy_load_test_enabled": false, "telemetry": "disabled", "token": "tid=ded1d75350f66adcb3d0ab36e8e78c47;exp=1774026494;sku=monthly_subscriber_quota;proxy-ep=proxy.individual.githubcopilot.com;st=dotcom;chat=1;cit=1;malfil=1;editor_preview_features=1;agent_mode=1;agent_mode_auto_approval=1;mcp=1;ccr=1;8kp=1;ip=24.143.97.87;asn=AS11404:c4aa3540c35631171800a409724219dcc61939aa9a96c454721b3fe4be4e713c", "tracking_id": "ded1d75350f66adcb3d0ab36e8e78c47", "vsc_electron_fetcher_v2": false, "xcode": true, "xcode_chat": false}
|
||||
@@ -1,5 +1,5 @@
|
||||
{
|
||||
"token": "tid=ded1d75350f66adcb3d0ab36e8e78c47;exp=1773915000;sku=monthly_subscriber_quota;proxy-ep=proxy.individual.githubcopilot.com;st=dotcom;chat=1;cit=1;malfil=1;editor_preview_features=1;agent_mode=1;agent_mode_auto_approval=1;mcp=1;ccr=1;8kp=1;ip=24.143.97.87;asn=AS11404:5580ca073a27022a75be9352a8ccf3a782a7074ad9e5b5cb52cf5f05215d0598",
|
||||
"expiresAt": 1773915000000,
|
||||
"updatedAt": 1773913200783
|
||||
"token": "tid=ded1d75350f66adcb3d0ab36e8e78c47;exp=1774002448;sku=monthly_subscriber_quota;proxy-ep=proxy.individual.githubcopilot.com;st=dotcom;chat=1;cit=1;malfil=1;editor_preview_features=1;agent_mode=1;agent_mode_auto_approval=1;mcp=1;ccr=1;8kp=1;ip=24.143.97.87;asn=AS11404:2983af469950d5538b500a5af6dbd746b2ab0b39e0fc570437cbe7c474680287",
|
||||
"expiresAt": 1774002448000,
|
||||
"updatedAt": 1774000648981
|
||||
}
|
||||
|
||||
@@ -7,7 +7,7 @@
|
||||
"description": "15m homelab service monitoring with Telegram alerting",
|
||||
"enabled": true,
|
||||
"createdAtMs": 1772676352724,
|
||||
"updatedAtMs": 1773903614233,
|
||||
"updatedAtMs": 1773990073687,
|
||||
"schedule": {
|
||||
"kind": "cron",
|
||||
"expr": "0 */4 * * *",
|
||||
@@ -28,14 +28,15 @@
|
||||
"bestEffort": true
|
||||
},
|
||||
"state": {
|
||||
"nextRunAtMs": 1773918000000,
|
||||
"lastRunAtMs": 1773903600006,
|
||||
"lastRunStatus": "ok",
|
||||
"lastStatus": "ok",
|
||||
"lastDurationMs": 14227,
|
||||
"nextRunAtMs": 1774004400000,
|
||||
"lastRunAtMs": 1773990000007,
|
||||
"lastRunStatus": "error",
|
||||
"lastStatus": "error",
|
||||
"lastDurationMs": 73680,
|
||||
"lastDeliveryStatus": "delivered",
|
||||
"consecutiveErrors": 0,
|
||||
"lastDelivered": true
|
||||
"consecutiveErrors": 1,
|
||||
"lastDelivered": true,
|
||||
"lastError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model."
|
||||
},
|
||||
"agentId": "automation"
|
||||
},
|
||||
@@ -263,7 +264,7 @@
|
||||
"name": "litellm model sync",
|
||||
"enabled": true,
|
||||
"createdAtMs": 1773260226908,
|
||||
"updatedAtMs": 1773908231724,
|
||||
"updatedAtMs": 1773994671307,
|
||||
"schedule": {
|
||||
"kind": "every",
|
||||
"everyMs": 43200000,
|
||||
@@ -281,14 +282,16 @@
|
||||
"mode": "none"
|
||||
},
|
||||
"state": {
|
||||
"nextRunAtMs": 1773951427013,
|
||||
"lastRunAtMs": 1773908227013,
|
||||
"lastRunStatus": "ok",
|
||||
"lastStatus": "ok",
|
||||
"lastDurationMs": 4711,
|
||||
"nextRunAtMs": 1774037827032,
|
||||
"lastRunAtMs": 1773994627032,
|
||||
"lastRunStatus": "error",
|
||||
"lastStatus": "error",
|
||||
"lastDurationMs": 44275,
|
||||
"lastDeliveryStatus": "not-delivered",
|
||||
"consecutiveErrors": 0,
|
||||
"lastDelivered": false
|
||||
"consecutiveErrors": 1,
|
||||
"lastDelivered": false,
|
||||
"lastError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model.",
|
||||
"lastDeliveryError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model."
|
||||
}
|
||||
},
|
||||
{
|
||||
@@ -333,7 +336,7 @@
|
||||
"name": "memory-reindex",
|
||||
"enabled": true,
|
||||
"createdAtMs": 1773778927874,
|
||||
"updatedAtMs": 1773914405934,
|
||||
"updatedAtMs": 1773999634293,
|
||||
"schedule": {
|
||||
"kind": "cron",
|
||||
"expr": "*/5 * * * *"
|
||||
@@ -349,14 +352,16 @@
|
||||
"mode": "none"
|
||||
},
|
||||
"state": {
|
||||
"nextRunAtMs": 1773914700000,
|
||||
"lastRunAtMs": 1773914400006,
|
||||
"lastRunStatus": "ok",
|
||||
"lastStatus": "ok",
|
||||
"lastDurationMs": 5928,
|
||||
"lastDelivered": false,
|
||||
"nextRunAtMs": 1774003234293,
|
||||
"lastRunAtMs": 1773999575526,
|
||||
"lastRunStatus": "error",
|
||||
"lastStatus": "error",
|
||||
"lastDurationMs": 58767,
|
||||
"lastDeliveryStatus": "not-delivered",
|
||||
"consecutiveErrors": 0
|
||||
"consecutiveErrors": 9,
|
||||
"lastDelivered": false,
|
||||
"lastDeliveryError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model.",
|
||||
"lastError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model."
|
||||
}
|
||||
},
|
||||
{
|
||||
@@ -366,7 +371,7 @@
|
||||
"name": "obsidian-inbox-watcher",
|
||||
"enabled": true,
|
||||
"createdAtMs": 1773780426888,
|
||||
"updatedAtMs": 1773914409598,
|
||||
"updatedAtMs": 1773999681244,
|
||||
"schedule": {
|
||||
"kind": "cron",
|
||||
"expr": "*/5 * * * *"
|
||||
@@ -382,14 +387,16 @@
|
||||
"mode": "none"
|
||||
},
|
||||
"state": {
|
||||
"nextRunAtMs": 1773914700000,
|
||||
"lastRunAtMs": 1773914405934,
|
||||
"lastRunStatus": "ok",
|
||||
"lastStatus": "ok",
|
||||
"lastDurationMs": 3664,
|
||||
"lastDelivered": false,
|
||||
"nextRunAtMs": 1774003281244,
|
||||
"lastRunAtMs": 1773999640204,
|
||||
"lastRunStatus": "error",
|
||||
"lastStatus": "error",
|
||||
"lastDurationMs": 41040,
|
||||
"lastDeliveryStatus": "not-delivered",
|
||||
"consecutiveErrors": 0
|
||||
"consecutiveErrors": 9,
|
||||
"lastDelivered": false,
|
||||
"lastDeliveryError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model.",
|
||||
"lastError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model."
|
||||
}
|
||||
},
|
||||
{
|
||||
@@ -422,7 +429,7 @@
|
||||
"name": "Weekly recycling reminder",
|
||||
"enabled": true,
|
||||
"createdAtMs": 1773853175958,
|
||||
"updatedAtMs": 1773853175958,
|
||||
"updatedAtMs": 1773997233199,
|
||||
"schedule": {
|
||||
"kind": "cron",
|
||||
"expr": "0 2 * * 5",
|
||||
@@ -438,7 +445,14 @@
|
||||
"mode": "announce"
|
||||
},
|
||||
"state": {
|
||||
"nextRunAtMs": 1773997200000
|
||||
"nextRunAtMs": 1774602000000,
|
||||
"lastRunAtMs": 1773997200005,
|
||||
"lastRunStatus": "error",
|
||||
"lastStatus": "error",
|
||||
"lastDurationMs": 33194,
|
||||
"lastError": "Delivering to Telegram requires target <chatId>",
|
||||
"lastDeliveryStatus": "unknown",
|
||||
"consecutiveErrors": 1
|
||||
}
|
||||
},
|
||||
{
|
||||
@@ -538,6 +552,42 @@
|
||||
"state": {
|
||||
"nextRunAtMs": 1775577600000
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "74a0d97f-f12f-4f20-afca-8cb57aae87f2",
|
||||
"agentId": "automation",
|
||||
"sessionKey": "agent:main:tui-79537d43-e1d3-4744-a67b-f7b88f8c8e41",
|
||||
"name": "n8n IMAP watchdog",
|
||||
"description": "Bounce IMAP IDLE workflows if they haven't fired in >90 minutes",
|
||||
"enabled": true,
|
||||
"createdAtMs": 1773988759710,
|
||||
"updatedAtMs": 1774000911858,
|
||||
"schedule": {
|
||||
"kind": "cron",
|
||||
"expr": "*/30 * * * *"
|
||||
},
|
||||
"sessionTarget": "isolated",
|
||||
"wakeMode": "now",
|
||||
"payload": {
|
||||
"kind": "agentTurn",
|
||||
"message": "Run the n8n IMAP watchdog script:\n\nbash /home/openclaw/.openclaw/workspace/scripts/n8n-imap-watchdog.sh\n\nIf exit code is 0 (all OK), reply exactly: NO_REPLY\nIf exit code is 2 (workflows bounced), send a Telegram message to Will: ⚠️ n8n IMAP watchdog bounced one or more workflows (IMAP IDLE had stalled). Check that email triage is working again.\nIf any other error, send a concise alert to Will via Telegram.",
|
||||
"timeoutSeconds": 60
|
||||
},
|
||||
"delivery": {
|
||||
"mode": "none"
|
||||
},
|
||||
"state": {
|
||||
"nextRunAtMs": 1774004511858,
|
||||
"lastRunAtMs": 1774000874373,
|
||||
"lastRunStatus": "error",
|
||||
"lastStatus": "error",
|
||||
"lastDurationMs": 37485,
|
||||
"lastDelivered": false,
|
||||
"lastDeliveryStatus": "not-delivered",
|
||||
"lastDeliveryError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model.",
|
||||
"consecutiveErrors": 6,
|
||||
"lastError": "Context overflow: prompt too large for the model. Try /reset (or /new) to start a fresh session, or use a larger-context model."
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
@@ -12,7 +12,7 @@
|
||||
"operator.read",
|
||||
"operator.write"
|
||||
],
|
||||
"updatedAtMs": 1773903605133
|
||||
"updatedAtMs": 1773989281449
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
{
|
||||
"version": 2,
|
||||
"lastUpdateId": 148910986,
|
||||
"lastUpdateId": 148910988,
|
||||
"botId": "8792219052"
|
||||
}
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
{
|
||||
"lastCheckedAt": "2026-03-18T04:26:50.033Z",
|
||||
"lastCheckedAt": "2026-03-19T18:20:40.956Z",
|
||||
"lastNotifiedVersion": "2026.3.13",
|
||||
"lastNotifiedTag": "latest"
|
||||
}
|
||||
Reference in New Issue
Block a user