26 Commits

Author SHA1 Message Date
7a3a28d87f add 0.6.4 2026-01-18 16:58:08 +00:00
240cded8a9 0.6.4 2026-01-18 16:56:23 +00:00
31e241f052 0.6.4 2026-01-18 16:56:04 +00:00
de579682a0 0.6.4 2026-01-18 16:55:36 +00:00
9acbd5046c Add 0.6.3 2026-01-18 15:54:21 +00:00
8d63c88e69 0.6.3 2026-01-18 15:53:52 +00:00
cffb0af60e 0.6.3 2026-01-18 15:53:03 +00:00
857b7a127a 0.6 3 2026-01-18 15:52:28 +00:00
66b24ece48 0.6.3 2026-01-18 15:52:01 +00:00
0cc3b466e0 0.6.3 2026-01-18 15:51:34 +00:00
f1e03b31a1 add 0.6.2 2026-01-18 13:16:42 +00:00
4e12d596d6 0.6.2 2026-01-18 13:12:47 +00:00
fa97f89afb 0.6.2 2026-01-18 13:12:21 +00:00
0718bee185 0.6.2 2026-01-18 13:11:39 +00:00
1a53107450 0.6.2 2026-01-18 13:10:59 +00:00
ab82cc6fd3 0.6.2 2026-01-18 13:10:16 +00:00
8e51f144e1 0.6.2 2026-01-18 13:09:37 +00:00
f292e22301 Fix restore version 2026-01-18 09:08:25 +00:00
2eb194c001 Add 0.6 1 2026-01-18 09:07:40 +00:00
f4e367987a 0.6.1 2026-01-18 09:07:03 +00:00
08aa4b5e15 0.6.0 2026-01-18 08:37:07 +00:00
b1676482f0 0.6.0 2026-01-18 08:36:34 +00:00
e46cd6e488 0.6.0 2026-01-18 08:34:44 +00:00
edd2fdd3fb 0.6.0 2026-01-18 08:33:34 +00:00
a4a0c1462b 0.6.0 2026-01-18 08:32:51 +00:00
196e63c08e 0.6.0 2026-01-18 08:32:06 +00:00
7 changed files with 1338 additions and 26 deletions

View File

@@ -11,6 +11,53 @@ Sections:
--- ---
## [0.6.4] - 2026-01-18
### Fixed
- Fixed long Home Assistant startup times caused by background repository enrichment running too early.
### Changed
- Background repository enrichment is now started only after Home Assistant has fully started.
- Repository cache updates now run fully asynchronous without blocking Home Assistant startup.
### Internal
- Improved alignment with Home Assistant startup lifecycle.
- No functional changes to store behavior or UI.
## [0.6.3] - 2026-01-18
### Changed
- Improved Store performance for large indexes by avoiding full metadata enrichment during list refresh.
- Repository details are loaded on demand, reducing initial load time and network requests.
- Index refresh is skipped when the index content has not changed.
## [0.6.2] - 2026-01-18
### Added
- Selectable install/update version per repository (install older releases/tags to downgrade when needed).
- New API endpoint to list available versions for a repository: `GET /api/bcs/versions?repo_id=...`.
## [0.6.1] - 2026-01-18
### Fixed
- Restored integrations now correctly report the restored version instead of the latest installed version.
- Update availability is correctly recalculated after restoring a backup, allowing updates to be applied again.
- Improved restore compatibility with backups created before version metadata was introduced.
## [0.6.0] - 2026-01-18
### Added
- Restore feature with selection of the last available backups (up to 5 per domain).
- New API endpoints to list and restore backups:
- `GET /api/bcs/backups?repo_id=...`
- `POST /api/bcs/restore?repo_id=...&backup_id=...`
### Safety
- Restoring a backup triggers a “restart required” prompt to apply the recovered integration state.
### Notes
- This is a major release milestone consolidating install/update/uninstall, backup/rollback, and restore workflows.
## [0.5.11] - 2026-01-18 ## [0.5.11] - 2026-01-18
### Added ### Added

View File

@@ -3,9 +3,10 @@ from __future__ import annotations
import logging import logging
from datetime import timedelta from datetime import timedelta
from homeassistant.core import HomeAssistant from homeassistant.core import HomeAssistant, callback
from homeassistant.const import EVENT_HOMEASSISTANT_STARTED
from homeassistant.components.panel_custom import async_register_panel from homeassistant.components.panel_custom import async_register_panel
from homeassistant.helpers.event import async_track_time_interval from homeassistant.helpers.event import async_track_time_interval, async_call_later
from homeassistant.helpers.discovery import async_load_platform from homeassistant.helpers.discovery import async_load_platform
from .core import BCSCore, BCSConfig, BCSError from .core import BCSCore, BCSConfig, BCSError
@@ -35,20 +36,28 @@ async def async_setup(hass: HomeAssistant, config: dict) -> bool:
StaticAssetsView, StaticAssetsView,
BCSApiView, BCSApiView,
BCSReadmeView, BCSReadmeView,
BCSVersionsView,
BCSRepoDetailView,
BCSCustomRepoView, BCSCustomRepoView,
BCSInstallView, BCSInstallView,
BCSUpdateView, BCSUpdateView,
BCSUninstallView, BCSUninstallView,
BCSBackupsView,
BCSRestoreView,
BCSRestartView, BCSRestartView,
) )
hass.http.register_view(StaticAssetsView()) hass.http.register_view(StaticAssetsView())
hass.http.register_view(BCSApiView(core)) hass.http.register_view(BCSApiView(core))
hass.http.register_view(BCSReadmeView(core)) hass.http.register_view(BCSReadmeView(core))
hass.http.register_view(BCSVersionsView(core))
hass.http.register_view(BCSRepoDetailView(core))
hass.http.register_view(BCSCustomRepoView(core)) hass.http.register_view(BCSCustomRepoView(core))
hass.http.register_view(BCSInstallView(core)) hass.http.register_view(BCSInstallView(core))
hass.http.register_view(BCSUpdateView(core)) hass.http.register_view(BCSUpdateView(core))
hass.http.register_view(BCSUninstallView(core)) hass.http.register_view(BCSUninstallView(core))
hass.http.register_view(BCSBackupsView(core))
hass.http.register_view(BCSRestoreView(core))
hass.http.register_view(BCSRestartView(core)) hass.http.register_view(BCSRestartView(core))
await async_register_panel( await async_register_panel(
@@ -56,17 +65,30 @@ async def async_setup(hass: HomeAssistant, config: dict) -> bool:
frontend_url_path="bahmcloud-store", frontend_url_path="bahmcloud-store",
webcomponent_name="bahmcloud-store-panel", webcomponent_name="bahmcloud-store-panel",
# IMPORTANT: bump v to avoid caching old JS # IMPORTANT: bump v to avoid caching old JS
module_url="/api/bahmcloud_store_static/panel.js?v=102", module_url="/api/bahmcloud_store_static/panel.js?v=107",
sidebar_title="Bahmcloud Store", sidebar_title="Bahmcloud Store",
sidebar_icon="mdi:store", sidebar_icon="mdi:store",
require_admin=True, require_admin=True,
config={}, config={},
) )
# IMPORTANT:
# Do NOT block Home Assistant startup with network-heavy refreshes.
# We wait until HA has fully started, then kick off the initial refresh.
async def _startup_refresh() -> None:
try: try:
await core.full_refresh(source="startup") await core.full_refresh(source="startup")
except BCSError as e: except BCSError as e:
_LOGGER.error("Initial refresh failed: %s", e) _LOGGER.error("Initial refresh failed: %s", e)
except Exception:
_LOGGER.exception("Unexpected error during initial refresh")
@callback
def _schedule_startup_refresh(_event=None) -> None:
# Give HA a short head-start (UI, recorder, etc.) before we start fetching lots of data.
async_call_later(hass, 30, lambda _now: hass.async_create_task(_startup_refresh()))
hass.bus.async_listen_once(EVENT_HOMEASSISTANT_STARTED, _schedule_startup_refresh)
async def periodic(_now) -> None: async def periodic(_now) -> None:
try: try:

View File

@@ -20,7 +20,7 @@ from homeassistant.helpers import issue_registry as ir
from homeassistant.util import yaml as ha_yaml from homeassistant.util import yaml as ha_yaml
from .storage import BCSStorage, CustomRepo from .storage import BCSStorage, CustomRepo
from .providers import fetch_repo_info, detect_provider, RepoInfo, fetch_readme_markdown from .providers import fetch_repo_info, detect_provider, RepoInfo, fetch_readme_markdown, fetch_repo_versions
from .metadata import fetch_repo_metadata, RepoMetadata from .metadata import fetch_repo_metadata, RepoMetadata
_LOGGER = logging.getLogger(__name__) _LOGGER = logging.getLogger(__name__)
@@ -30,6 +30,8 @@ DOMAIN = "bahmcloud_store"
SIGNAL_UPDATED = f"{DOMAIN}_updated" SIGNAL_UPDATED = f"{DOMAIN}_updated"
RESTART_REQUIRED_ISSUE_ID = "restart_required" RESTART_REQUIRED_ISSUE_ID = "restart_required"
BACKUP_META_FILENAME = ".bcs_backup_meta.json"
class BCSError(Exception): class BCSError(Exception):
"""BCS core error.""" """BCS core error."""
@@ -87,9 +89,27 @@ class BCSCore:
self.last_index_hash: str | None = None self.last_index_hash: str | None = None
self.last_index_loaded_at: float | None = None self.last_index_loaded_at: float | None = None
# Fast refresh: skip expensive processing when index/custom repos unchanged
self._last_refresh_signature: str | None = None
self._install_lock = asyncio.Lock() self._install_lock = asyncio.Lock()
self._installed_cache: dict[str, Any] = {} self._installed_cache: dict[str, Any] = {}
# Phase P1/P2: local repo cache + background enrichment
# The cache persists provider/meta/latest data so the UI can show more
# information immediately and we can later do delta refresh.
self._repo_cache: dict[str, Any] = {}
self._repo_cache_loaded: bool = False
self._repo_cache_flush_task: asyncio.Task | None = None
# Background enrichment worker (non-blocking)
self._bg_enrich_task: asyncio.Task | None = None
self._bg_enrich_pending: set[str] = set()
self._bg_enrich_ttl_seconds: int = 6 * 3600
self._bg_enrich_max_parallel: int = 3
self._bg_signal_interval_seconds: float = 2.0
self._bg_last_signal_ts: float = 0.0
# Phase F2: backups before install/update # Phase F2: backups before install/update
self._backup_root = Path(self.hass.config.path(".bcs_backups")) self._backup_root = Path(self.hass.config.path(".bcs_backups"))
self._backup_keep_per_domain: int = 5 self._backup_keep_per_domain: int = 5
@@ -99,6 +119,15 @@ class BCSCore:
self.version = await self._read_manifest_version_async() self.version = await self._read_manifest_version_async()
await self._refresh_installed_cache() await self._refresh_installed_cache()
# Load persisted repo cache once at startup.
try:
self._repo_cache = await self.storage.get_repo_cache_map()
if not isinstance(self._repo_cache, dict):
self._repo_cache = {}
except Exception:
self._repo_cache = {}
self._repo_cache_loaded = True
# After a successful HA restart, restart-required is no longer relevant. # After a successful HA restart, restart-required is no longer relevant.
self._clear_restart_required_issue() self._clear_restart_required_issue()
@@ -176,6 +205,19 @@ class BCSCore:
custom_repos = await self.storage.list_custom_repos() custom_repos = await self.storage.list_custom_repos()
# Fast path: if index + custom repos did not change, skip expensive work.
try:
custom_sig = [(c.id, (c.url or '').strip(), (c.name or '').strip()) for c in (custom_repos or [])]
custom_sig.sort()
refresh_signature = json.dumps({"index_hash": self.last_index_hash, "custom": custom_sig}, sort_keys=True)
except Exception:
refresh_signature = f"{self.last_index_hash}:{len(custom_repos or [])}"
if self._last_refresh_signature and refresh_signature == self._last_refresh_signature and self.repos:
_LOGGER.debug("BCS refresh skipped (no changes detected)")
return
merged: dict[str, RepoItem] = {} merged: dict[str, RepoItem] = {}
for item in index_repos: for item in index_repos:
@@ -192,9 +234,15 @@ class BCSCore:
for r in merged.values(): for r in merged.values():
r.provider = detect_provider(r.url) r.provider = detect_provider(r.url)
await self._enrich_and_resolve(merged) # Apply persisted cache (provider/meta/latest) to all repos so the list
# view can show richer data immediately.
self._apply_repo_cache(merged)
await self._enrich_installed_only(merged)
self.repos = merged self.repos = merged
self._last_refresh_signature = refresh_signature
_LOGGER.info( _LOGGER.info(
"BCS refresh complete: repos=%s (index=%s, custom=%s)", "BCS refresh complete: repos=%s (index=%s, custom=%s)",
len(self.repos), len(self.repos),
@@ -202,6 +250,159 @@ class BCSCore:
len([r for r in self.repos.values() if r.source == "custom"]), len([r for r in self.repos.values() if r.source == "custom"]),
) )
# Start/continue background enrichment for repos (non-blocking).
self._schedule_background_enrich(list(self.repos.keys()))
def _apply_repo_cache(self, merged: dict[str, RepoItem]) -> None:
"""Apply persisted cached enrichment data to repo items (no network IO)."""
if not self._repo_cache_loaded or not isinstance(self._repo_cache, dict) or not self._repo_cache:
return
for rid, r in merged.items():
entry = self._repo_cache.get(str(rid))
if not isinstance(entry, dict):
continue
if (entry.get("url") or "").strip() != (r.url or "").strip():
continue
# Provider basics
r.provider = entry.get("provider") or r.provider
r.owner = entry.get("owner") or r.owner
r.provider_repo_name = entry.get("provider_repo_name") or r.provider_repo_name
r.provider_description = entry.get("provider_description") or r.provider_description
r.default_branch = entry.get("default_branch") or r.default_branch
# Latest version
r.latest_version = entry.get("latest_version") or r.latest_version
r.latest_version_source = entry.get("latest_version_source") or r.latest_version_source
# Metadata
r.meta_source = entry.get("meta_source") or r.meta_source
r.meta_name = entry.get("meta_name") or r.meta_name
r.meta_description = entry.get("meta_description") or r.meta_description
r.meta_category = entry.get("meta_category") or r.meta_category
r.meta_author = entry.get("meta_author") or r.meta_author
r.meta_maintainer = entry.get("meta_maintainer") or r.meta_maintainer
# Keep a stable name fallback
if r.meta_name:
r.name = r.meta_name
elif not r.name:
r.name = r.provider_repo_name or r.url
def _schedule_repo_cache_flush(self) -> None:
if self._repo_cache_flush_task and not self._repo_cache_flush_task.done():
return
async def _flush_delayed() -> None:
await asyncio.sleep(5)
try:
await self.storage.set_repo_cache_map(self._repo_cache)
except Exception:
_LOGGER.debug("BCS repo cache flush failed", exc_info=True)
self._repo_cache_flush_task = self.hass.async_create_task(_flush_delayed())
def _cache_entry_is_stale(self, entry: dict[str, Any]) -> bool:
try:
checked_at = int(entry.get("checked_at") or 0)
except Exception:
checked_at = 0
if checked_at <= 0:
return True
return (time.time() - checked_at) > self._bg_enrich_ttl_seconds
def _schedule_background_enrich(self, repo_ids: list[str]) -> None:
"""Queue repos for background enrichment and ensure worker is running."""
if not repo_ids:
return
now = time.time()
for rid in repo_ids:
rid = str(rid)
r = self.repos.get(rid)
if not r:
continue
# Already enriched in memory? Still consider staleness from cache.
entry = self._repo_cache.get(rid) if isinstance(self._repo_cache, dict) else None
stale = True
if isinstance(entry, dict) and (entry.get("url") or "").strip() == (r.url or "").strip():
stale = self._cache_entry_is_stale(entry)
# If we already have fields in memory and the cache isn't stale, skip.
if (r.latest_version or r.meta_source or r.provider_description) and not stale:
continue
# Always enqueue missing/stale entries.
self._bg_enrich_pending.add(rid)
if self._bg_enrich_task and not self._bg_enrich_task.done():
return
self._bg_enrich_task = self.hass.async_create_task(self._background_enrich_worker())
async def _background_enrich_worker(self) -> None:
"""Background worker to enrich repos and update the persistent cache."""
sem = asyncio.Semaphore(self._bg_enrich_max_parallel)
async def _enrich_one(rid: str) -> None:
async with sem:
r = self.repos.get(rid)
if not r:
return
entry = self._repo_cache.get(rid) if isinstance(self._repo_cache, dict) else None
if isinstance(entry, dict) and (entry.get("url") or "").strip() == (r.url or "").strip():
if not self._cache_entry_is_stale(entry) and (r.latest_version or r.meta_source or r.provider_description):
return
try:
await self._enrich_one_repo(r)
except Exception:
_LOGGER.debug("BCS background enrich failed for %s", rid, exc_info=True)
# still mark checked_at to avoid tight retry loops
self._repo_cache[rid] = {
"url": r.url,
"checked_at": int(time.time()),
}
self._schedule_repo_cache_flush()
return
# Update persistent cache entry
self._repo_cache[rid] = {
"url": r.url,
"provider": r.provider,
"owner": r.owner,
"provider_repo_name": r.provider_repo_name,
"provider_description": r.provider_description,
"default_branch": r.default_branch,
"latest_version": r.latest_version,
"latest_version_source": r.latest_version_source,
"meta_source": r.meta_source,
"meta_name": r.meta_name,
"meta_description": r.meta_description,
"meta_category": r.meta_category,
"meta_author": r.meta_author,
"meta_maintainer": r.meta_maintainer,
"checked_at": int(time.time()),
}
self._schedule_repo_cache_flush()
# Throttle UI/entity updates
if (time.time() - self._bg_last_signal_ts) >= self._bg_signal_interval_seconds:
self._bg_last_signal_ts = time.time()
self.signal_updated()
while self._bg_enrich_pending:
# Drain in small batches so we don't monopolize the loop
batch: list[str] = []
while self._bg_enrich_pending and len(batch) < (self._bg_enrich_max_parallel * 2):
batch.append(self._bg_enrich_pending.pop())
await asyncio.gather(*(_enrich_one(rid) for rid in batch), return_exceptions=True)
await asyncio.sleep(0) # yield
async def _enrich_and_resolve(self, merged: dict[str, RepoItem]) -> None: async def _enrich_and_resolve(self, merged: dict[str, RepoItem]) -> None:
sem = asyncio.Semaphore(6) sem = asyncio.Semaphore(6)
@@ -236,6 +437,178 @@ class BCSCore:
await asyncio.gather(*(process_one(r) for r in merged.values()), return_exceptions=True) await asyncio.gather(*(process_one(r) for r in merged.values()), return_exceptions=True)
async def _enrich_installed_only(self, merged: dict[str, RepoItem]) -> None:
"""Enrich only installed repos (fast refresh for large indexes).
This keeps the backend responsive even with thousands of repositories.
Details for non-installed repos are fetched on-demand.
"""
installed_map: dict[str, Any] = getattr(self, "_installed_cache", {}) or {}
if not isinstance(installed_map, dict) or not installed_map:
return
to_process: list[RepoItem] = []
for rid in installed_map.keys():
r = merged.get(str(rid))
if r:
to_process.append(r)
if not to_process:
return
sem = asyncio.Semaphore(6)
async def process_one(r: RepoItem) -> None:
async with sem:
await self._enrich_one_repo(r)
await asyncio.gather(*(process_one(r) for r in to_process), return_exceptions=True)
async def _enrich_one_repo(self, r: RepoItem) -> None:
"""Fetch provider info + metadata for a single repo item."""
info: RepoInfo = await fetch_repo_info(self.hass, r.url)
r.provider = info.provider or r.provider
r.owner = info.owner or r.owner
r.provider_repo_name = info.repo_name
r.provider_description = info.description
r.default_branch = info.default_branch or r.default_branch
r.latest_version = info.latest_version
r.latest_version_source = info.latest_version_source
md: RepoMetadata = await fetch_repo_metadata(self.hass, r.url, r.default_branch)
r.meta_source = md.source
if md.name:
r.meta_name = md.name
r.name = md.name
r.meta_description = md.description
if md.category:
r.meta_category = md.category
r.meta_author = md.author
r.meta_maintainer = md.maintainer
# Keep a stable name fallback
if not r.name:
r.name = r.provider_repo_name or r.url
# Persist into local cache (non-blocking, throttled flush).
self._update_repo_cache_from_item(r)
async def ensure_repo_details(self, repo_id: str) -> RepoItem | None:
"""Ensure provider/meta/latest fields are loaded for a repo.
Used by the UI when a repo detail view is opened.
"""
r = self.get_repo(repo_id)
if not r:
return None
# If we already have a latest_version (or provider_description), consider it enriched.
if r.latest_version or r.provider_description or r.meta_source:
return r
try:
await self._enrich_one_repo(r)
except Exception:
_LOGGER.debug("BCS ensure_repo_details failed for %s", repo_id, exc_info=True)
return r
# ---------------------------------------------------------------------
# Phase P1/P2: local cache + background enrichment
# ---------------------------------------------------------------------
def _apply_repo_cache(self, merged: dict[str, RepoItem]) -> None:
"""Apply persisted cache fields to repo items.
This makes the list view richer immediately (without remote requests).
"""
cache = self._repo_cache if isinstance(self._repo_cache, dict) else {}
now = int(time.time())
for rid, r in merged.items():
entry = cache.get(str(rid))
if not isinstance(entry, dict):
continue
# Safety: ensure cache belongs to the same URL.
if str(entry.get("url") or "").strip() != str(r.url or "").strip():
continue
# Provider fields
r.provider = entry.get("provider") or r.provider
r.owner = entry.get("owner") or r.owner
r.provider_repo_name = entry.get("provider_repo_name") or r.provider_repo_name
r.provider_description = entry.get("provider_description") or r.provider_description
r.default_branch = entry.get("default_branch") or r.default_branch
# Latest version
r.latest_version = entry.get("latest_version") or r.latest_version
r.latest_version_source = entry.get("latest_version_source") or r.latest_version_source
# Metadata
r.meta_source = entry.get("meta_source") or r.meta_source
r.meta_name = entry.get("meta_name") or r.meta_name
r.meta_description = entry.get("meta_description") or r.meta_description
r.meta_category = entry.get("meta_category") or r.meta_category
r.meta_author = entry.get("meta_author") or r.meta_author
r.meta_maintainer = entry.get("meta_maintainer") or r.meta_maintainer
# Stable display name
if r.meta_name:
r.name = r.meta_name
elif not r.name:
r.name = r.provider_repo_name or r.url
# Mark as stale if the cache is old (used by background enrich).
checked_at = int(entry.get("checked_at") or 0)
entry["_stale"] = (checked_at <= 0) or ((now - checked_at) > self._bg_enrich_ttl_seconds)
def _update_repo_cache_from_item(self, r: RepoItem) -> None:
"""Update in-memory cache from a repo item and schedule a flush."""
if not self._repo_cache_loaded:
return
rid = str(r.id)
now = int(time.time())
entry = {
"url": str(r.url or ""),
"provider": r.provider,
"owner": r.owner,
"provider_repo_name": r.provider_repo_name,
"provider_description": r.provider_description,
"default_branch": r.default_branch,
"latest_version": r.latest_version,
"latest_version_source": r.latest_version_source,
"meta_source": r.meta_source,
"meta_name": r.meta_name,
"meta_description": r.meta_description,
"meta_category": r.meta_category,
"meta_author": r.meta_author,
"meta_maintainer": r.meta_maintainer,
"checked_at": now,
}
if not isinstance(self._repo_cache, dict):
self._repo_cache = {}
self._repo_cache[rid] = entry
self._schedule_repo_cache_flush()
def _schedule_repo_cache_flush(self) -> None:
if self._repo_cache_flush_task and not self._repo_cache_flush_task.done():
return
async def _flush_later() -> None:
try:
await asyncio.sleep(5)
if isinstance(self._repo_cache, dict):
await self.storage.set_repo_cache_map(self._repo_cache)
except Exception:
_LOGGER.debug("BCS repo cache flush failed", exc_info=True)
self._repo_cache_flush_task = self.hass.async_create_task(_flush_later())
def _add_cache_buster(self, url: str) -> str: def _add_cache_buster(self, url: str) -> str:
parts = urlsplit(url) parts = urlsplit(url)
q = dict(parse_qsl(parts.query, keep_blank_values=True)) q = dict(parse_qsl(parts.query, keep_blank_values=True))
@@ -327,6 +700,7 @@ class BCSCore:
name=name, name=name,
url=repo_url, url=repo_url,
source="index", source="index",
meta_category=str(r.get("category")) if r.get("category") else None,
) )
) )
@@ -413,6 +787,23 @@ class BCSCore:
default_branch=repo.default_branch, default_branch=repo.default_branch,
) )
async def list_repo_versions(self, repo_id: str, *, limit: int = 20) -> list[dict[str, str]]:
"""List installable versions/refs for a repo.
This is used by the UI to allow selecting an older tag/release.
"""
repo = self.get_repo(repo_id)
if not repo:
raise BCSInstallError(f"repo_id not found: {repo_id}")
return await fetch_repo_versions(
self.hass,
repo.url,
provider=repo.provider,
default_branch=repo.default_branch,
limit=limit,
)
def _pick_ref_for_install(self, repo: RepoItem) -> str: def _pick_ref_for_install(self, repo: RepoItem) -> str:
if repo.latest_version and str(repo.latest_version).strip(): if repo.latest_version and str(repo.latest_version).strip():
return str(repo.latest_version).strip() return str(repo.latest_version).strip()
@@ -493,7 +884,19 @@ class BCSCore:
await self.hass.async_add_executor_job(_mkdir) await self.hass.async_add_executor_job(_mkdir)
async def _backup_domain(self, domain: str) -> Path | None: def _build_backup_meta(self, repo_id: str, domain: str) -> dict[str, object]:
"""Build metadata for backup folders so restores can recover the stored version."""
inst = self.get_installed(repo_id) or {}
return {
"repo_id": repo_id,
"domain": domain,
"installed_version": inst.get("installed_version") or inst.get("ref"),
"installed_manifest_version": inst.get("installed_manifest_version"),
"ref": inst.get("ref") or inst.get("installed_version"),
"created_at": time.strftime("%Y-%m-%dT%H:%M:%S"),
}
async def _backup_domain(self, domain: str, *, meta: dict[str, object] | None = None) -> Path | None:
"""Backup an existing domain folder. """Backup an existing domain folder.
Returns the created backup path, or None if the domain folder does not exist. Returns the created backup path, or None if the domain folder does not exist.
@@ -515,6 +918,13 @@ class BCSCore:
if backup_path.exists(): if backup_path.exists():
shutil.rmtree(backup_path, ignore_errors=True) shutil.rmtree(backup_path, ignore_errors=True)
shutil.copytree(target, backup_path, dirs_exist_ok=True) shutil.copytree(target, backup_path, dirs_exist_ok=True)
# Store backup metadata (kept inside backup folder; removed from target after restore).
if meta:
try:
meta_path = backup_path / BACKUP_META_FILENAME
meta_path.write_text(json.dumps(meta, ensure_ascii=False, indent=2), encoding='utf-8')
except Exception:
pass
# Retention: keep only the newest N backups per domain. # Retention: keep only the newest N backups per domain.
try: try:
@@ -541,10 +951,237 @@ class BCSCore:
if target.exists(): if target.exists():
shutil.rmtree(target, ignore_errors=True) shutil.rmtree(target, ignore_errors=True)
shutil.copytree(backup_path, target, dirs_exist_ok=True) shutil.copytree(backup_path, target, dirs_exist_ok=True)
try:
meta_file = target / BACKUP_META_FILENAME
if meta_file.exists():
meta_file.unlink(missing_ok=True)
except Exception:
pass
# Do not leave backup metadata inside the restored integration folder.
try:
meta_p = target / BACKUP_META_FILENAME
if meta_p.exists():
meta_p.unlink()
except Exception:
pass
# Do not leave backup metadata inside the live integration folder.
try:
m = target / BACKUP_META_FILENAME
if m.exists():
m.unlink()
except Exception:
pass
await self.hass.async_add_executor_job(_restore) await self.hass.async_add_executor_job(_restore)
_LOGGER.info("BCS rollback applied: domain=%s from=%s", domain, backup_path) _LOGGER.info("BCS rollback applied: domain=%s from=%s", domain, backup_path)
async def list_repo_backups(self, repo_id: str) -> list[dict[str, Any]]:
"""List available backup sets for an installed repository.
Returns a list of items sorted newest->oldest:
{"id": "YYYYMMDD_HHMMSS", "label": "YYYY-MM-DD HH:MM:SS", "complete": bool, "domains": [...] }
A backup set is considered *complete* if the timestamp exists for all
domains of the repository.
"""
inst = self.get_installed(repo_id) or {}
domains = inst.get("domains") or []
if not isinstance(domains, list) or not domains:
return []
dom_list = [str(d) for d in domains if str(d).strip()]
if not dom_list:
return []
# Collect timestamps per domain.
per_domain: dict[str, list[str]] = {}
for d in dom_list:
per_domain[d] = await self._list_domain_backup_ids(d)
# Build a map id -> domains where it exists
id_map: dict[str, set[str]] = {}
for d, ids in per_domain.items():
for bid in ids:
id_map.setdefault(bid, set()).add(d)
all_domains = set(dom_list)
items: list[dict[str, Any]] = []
for bid, present in id_map.items():
complete = present == all_domains
label = self._format_backup_id(bid)
meta = await self._read_backup_meta(dom_list[0], bid)
ver = None
if isinstance(meta, dict):
ver = meta.get("installed_version") or meta.get("ref")
if ver:
label = f"{label} ({ver})"
items.append({"id": bid, "label": label, "complete": complete, "domains": sorted(present), "installed_version": str(ver) if ver else None})
# Sort newest first by id (lexicographic works for timestamp format).
items.sort(key=lambda x: str(x.get("id") or ""), reverse=True)
# Keep newest 5 entries overall (UI expects up to 5).
return items[: self._backup_keep_per_domain]
async def restore_repo_backup(self, repo_id: str, backup_id: str) -> dict[str, Any]:
"""Restore a previously created backup set for a repository."""
repo_id = str(repo_id or "").strip()
backup_id = str(backup_id or "").strip()
if not repo_id:
raise BCSInstallError("Missing repo_id")
if not backup_id:
raise BCSInstallError("Missing backup_id")
inst = self.get_installed(repo_id)
if not inst:
raise BCSInstallError("Repository is not installed")
domains = inst.get("domains") or []
if not isinstance(domains, list) or not domains:
raise BCSInstallError("No installed domains found")
dom_list = [str(d) for d in domains if str(d).strip()]
if not dom_list:
raise BCSInstallError("No installed domains found")
# Ensure the backup exists for all domains.
missing: list[str] = []
for d in dom_list:
p = self._backup_root / d / backup_id
if not p.exists() or not p.is_dir():
missing.append(d)
if missing:
raise BCSInstallError(f"Selected backup is not available for all domains: missing={missing}")
async with self._install_lock:
_LOGGER.info("BCS restore started: repo_id=%s backup_id=%s domains=%s", repo_id, backup_id, dom_list)
# Safety: create a new backup of current state before restoring.
for d in dom_list:
try:
await self._backup_domain(d)
except Exception:
_LOGGER.debug("BCS pre-restore backup failed for domain=%s", d, exc_info=True)
# Apply restore.
for d in dom_list:
await self._restore_domain_from_backup(d, self._backup_root / d / backup_id)
# Update stored installed version to the restored one (so UI shows the restored state).
#
# Backups created before 0.6.1 may not have metadata. For those legacy backups we fall back to:
# 1) version from the backup's manifest.json (best-effort), else
# 2) a synthetic marker (restored:<backup_id>) so the UI reflects a restored state and updates
# remain available.
restored_meta = await self._read_backup_meta(dom_list[0], backup_id)
restored_version: str | None = None
restored_manifest_version: str | None = None
if isinstance(restored_meta, dict):
rv = restored_meta.get("installed_version") or restored_meta.get("ref")
if rv is not None and str(rv).strip():
restored_version = str(rv).strip()
mv = restored_meta.get("installed_manifest_version")
if mv is not None and str(mv).strip():
restored_manifest_version = str(mv).strip()
# Legacy backups (no meta): try to read manifest.json version from the backup folder.
if not restored_manifest_version:
restored_manifest_version = await self._read_backup_manifest_version(dom_list[0], backup_id)
# Use manifest version as a fallback display value if we don't have the exact installed ref.
if not restored_version and restored_manifest_version:
restored_version = restored_manifest_version
# Last resort: ensure the installed version changes so the UI does not keep showing the newest version.
if not restored_version:
restored_version = f"restored:{backup_id}"
repo = self.get_repo(repo_id)
repo_url = getattr(repo, "url", None) or ""
await self.storage.set_installed_repo(
repo_id=repo_id,
url=repo_url,
domains=dom_list,
installed_version=restored_version,
installed_manifest_version=restored_manifest_version,
ref=restored_version,
)
await self._refresh_installed_cache()
self._mark_restart_required()
self.signal_updated()
_LOGGER.info("BCS restore complete: repo_id=%s backup_id=%s", repo_id, backup_id)
return {"ok": True, "repo_id": repo_id, "backup_id": backup_id, "domains": dom_list, "restored_version": restored_version, "restart_required": True}
async def _read_backup_meta(self, domain: str, backup_id: str) -> dict[str, Any] | None:
"""Read backup metadata for a domain backup.
Metadata is stored inside the backup folder and will be removed from the
live folder after restore.
"""
try:
p = self._backup_root / domain / backup_id / BACKUP_META_FILENAME
if not p.exists():
return None
txt = await self.hass.async_add_executor_job(p.read_text, 'utf-8')
data = json.loads(txt)
return data if isinstance(data, dict) else None
except Exception:
return None
async def _read_backup_manifest_version(self, domain: str, backup_id: str) -> str | None:
"""Best-effort: read manifest.json version from a legacy backup (no metadata)."""
def _read() -> str | None:
try:
p = self._backup_root / domain / backup_id / 'manifest.json'
if not p.exists():
return None
data = json.loads(p.read_text(encoding='utf-8'))
v = data.get('version')
return str(v) if v else None
except Exception:
return None
return await self.hass.async_add_executor_job(_read)
async def _list_domain_backup_ids(self, domain: str) -> list[str]:
"""List backup ids for a domain (newest->oldest)."""
domain = str(domain or "").strip()
if not domain:
return []
root = self._backup_root / domain
def _list() -> list[str]:
if not root.exists() or not root.is_dir():
return []
ids = [p.name for p in root.iterdir() if p.is_dir()]
ids.sort(reverse=True)
return ids
ids = await self.hass.async_add_executor_job(_list)
return ids[: self._backup_keep_per_domain]
@staticmethod
def _format_backup_id(backup_id: str) -> str:
"""Format backup id YYYYMMDD_HHMMSS -> YYYY-MM-DD HH:MM:SS."""
s = str(backup_id or "").strip()
if len(s) != 15 or "_" not in s:
return s
try:
d, t = s.split("_", 1)
return f"{d[0:4]}-{d[4:6]}-{d[6:8]} {t[0:2]}:{t[2:4]}:{t[4:6]}"
except Exception:
return s
async def _copy_domain_dir(self, src_domain_dir: Path, domain: str) -> None: async def _copy_domain_dir(self, src_domain_dir: Path, domain: str) -> None:
dest_root = Path(self.hass.config.path("custom_components")) dest_root = Path(self.hass.config.path("custom_components"))
target = dest_root / domain target = dest_root / domain
@@ -665,19 +1302,28 @@ class BCSCore:
self.signal_updated() self.signal_updated()
return {"ok": True, "repo_id": repo_id, "removed": removed, "restart_required": bool(removed)} return {"ok": True, "repo_id": repo_id, "removed": removed, "restart_required": bool(removed)}
async def install_repo(self, repo_id: str) -> dict[str, Any]: async def install_repo(self, repo_id: str, *, version: str | None = None) -> dict[str, Any]:
repo = self.get_repo(repo_id) repo = self.get_repo(repo_id)
if not repo: if not repo:
raise BCSInstallError(f"repo_id not found: {repo_id}") raise BCSInstallError(f"repo_id not found: {repo_id}")
async with self._install_lock: async with self._install_lock:
ref = self._pick_ref_for_install(repo) requested = (version or "").strip()
ref = requested if requested else self._pick_ref_for_install(repo)
zip_url = self._build_zip_url(repo.url, ref) zip_url = self._build_zip_url(repo.url, ref)
_LOGGER.info("BCS install started: repo_id=%s ref=%s zip_url=%s", repo_id, ref, zip_url) _LOGGER.info("BCS install started: repo_id=%s ref=%s zip_url=%s", repo_id, ref, zip_url)
installed_domains: list[str] = [] installed_domains: list[str] = []
backups: dict[str, Path] = {} backups: dict[str, Path] = {}
inst_before = self.get_installed(repo_id) or {}
backup_meta = {
"repo_id": repo_id,
"installed_version": inst_before.get("installed_version") or inst_before.get("ref"),
"installed_manifest_version": inst_before.get("installed_manifest_version"),
"ref": inst_before.get("ref") or inst_before.get("installed_version"),
}
created_new: set[str] = set() created_new: set[str] = set()
try: try:
@@ -708,7 +1354,9 @@ class BCSCore:
# Backup only if we are going to overwrite an existing domain. # Backup only if we are going to overwrite an existing domain.
if target.exists() and target.is_dir(): if target.exists() and target.is_dir():
bkp = await self._backup_domain(domain) m = dict(backup_meta)
m["domain"] = domain
bkp = await self._backup_domain(domain, meta=m)
if bkp: if bkp:
backups[domain] = bkp backups[domain] = bkp
else: else:
@@ -781,9 +1429,9 @@ class BCSCore:
raise raise
raise BCSInstallError(str(e)) from e raise BCSInstallError(str(e)) from e
async def update_repo(self, repo_id: str) -> dict[str, Any]: async def update_repo(self, repo_id: str, *, version: str | None = None) -> dict[str, Any]:
_LOGGER.info("BCS update started: repo_id=%s", repo_id) _LOGGER.info("BCS update started: repo_id=%s", repo_id)
return await self.install_repo(repo_id) return await self.install_repo(repo_id, version=version)
async def request_restart(self) -> None: async def request_restart(self) -> None:
await self.hass.services.async_call("homeassistant", "restart", {}, blocking=False) await self.hass.services.async_call("homeassistant", "restart", {}, blocking=False)

View File

@@ -1,7 +1,7 @@
{ {
"domain": "bahmcloud_store", "domain": "bahmcloud_store",
"name": "Bahmcloud Store", "name": "Bahmcloud Store",
"version": "0.5.11", "version": "0.6.4",
"documentation": "https://git.bahmcloud.de/bahmcloud/bahmcloud_store", "documentation": "https://git.bahmcloud.de/bahmcloud/bahmcloud_store",
"platforms": ["update"], "platforms": ["update"],
"requirements": [], "requirements": [],

View File

@@ -36,6 +36,20 @@ class BahmcloudStorePanel extends HTMLElement {
this._uninstallingRepoId = null; this._uninstallingRepoId = null;
this._restartRequired = false; this._restartRequired = false;
this._lastActionMsg = null; this._lastActionMsg = null;
// Phase F2.1: restore from backups
this._restoreOpen = false;
this._restoreRepoId = null;
this._restoreLoading = false;
this._restoreOptions = [];
this._restoreSelected = "";
this._restoring = false;
this._restoreError = null;
// Phase C1: selectable install version
this._versionsCache = {}; // repo_id -> [{ref,label,source}, ...]
this._versionsLoadingRepoId = null;
this._selectedVersionByRepoId = {}; // repo_id -> ref ("" means latest)
} }
set hass(hass) { set hass(hass) {
@@ -105,7 +119,13 @@ class BahmcloudStorePanel extends HTMLElement {
this._update(); this._update();
try { try {
const resp = await this._hass.callApi("post", `bcs/install?repo_id=${encodeURIComponent(repoId)}`, {}); const sel = this._safeText(this._selectedVersionByRepoId?.[repoId] || "").trim();
const qv = sel ? `&version=${encodeURIComponent(sel)}` : "";
const resp = await this._hass.callApi(
"post",
`bcs/install?repo_id=${encodeURIComponent(repoId)}${qv}`,
{},
);
if (!resp?.ok) { if (!resp?.ok) {
this._error = this._safeText(resp?.message) || "Install failed."; this._error = this._safeText(resp?.message) || "Install failed.";
} else { } else {
@@ -131,7 +151,13 @@ class BahmcloudStorePanel extends HTMLElement {
this._update(); this._update();
try { try {
const resp = await this._hass.callApi("post", `bcs/update?repo_id=${encodeURIComponent(repoId)}`, {}); const sel = this._safeText(this._selectedVersionByRepoId?.[repoId] || "").trim();
const qv = sel ? `&version=${encodeURIComponent(sel)}` : "";
const resp = await this._hass.callApi(
"post",
`bcs/update?repo_id=${encodeURIComponent(repoId)}${qv}`,
{},
);
if (!resp?.ok) { if (!resp?.ok) {
this._error = this._safeText(resp?.message) || "Update failed."; this._error = this._safeText(resp?.message) || "Update failed.";
} else { } else {
@@ -175,6 +201,90 @@ class BahmcloudStorePanel extends HTMLElement {
} }
} }
async _openRestore(repoId) {
if (!this._hass) return;
if (!repoId) return;
if (this._installingRepoId || this._updatingRepoId || this._uninstallingRepoId) return;
this._restoreRepoId = repoId;
this._restoreOpen = true;
this._restoreLoading = true;
this._restoreError = null;
this._restoreOptions = [];
this._restoreSelected = "";
this._update();
try {
const resp = await this._hass.callApi("get", `bcs/backups?repo_id=${encodeURIComponent(repoId)}`);
const list = Array.isArray(resp?.backups) ? resp.backups : [];
this._restoreOptions = list;
const firstComplete = list.find((x) => x && x.complete);
const firstAny = list[0];
const pick = firstComplete || firstAny;
if (pick && pick.id) this._restoreSelected = String(pick.id);
if (!list.length) {
this._restoreError = "No backups found for this repository.";
}
} catch (e) {
this._restoreError = e?.message ? String(e.message) : String(e);
} finally {
this._restoreLoading = false;
this._update();
}
}
_closeRestore() {
this._restoreOpen = false;
this._restoreRepoId = null;
this._restoreLoading = false;
this._restoreError = null;
this._restoreOptions = [];
this._restoreSelected = "";
this._restoring = false;
this._update();
}
async _restoreSelectedBackup() {
if (!this._hass) return;
if (!this._restoreRepoId) return;
const bid = String(this._restoreSelected || "").trim();
if (!bid) return;
if (this._restoring) return;
const chosen = (this._restoreOptions || []).find((x) => String(x?.id) === bid);
if (chosen && chosen.complete === false) {
this._restoreError = "Selected backup is not available for all domains of this repository.";
this._update();
return;
}
const ok = window.confirm("Restore selected backup? This will overwrite the installed files under /config/custom_components and requires a restart.");
if (!ok) return;
this._restoring = true;
this._restoreError = null;
this._update();
try {
const resp = await this._hass.callApi("post", `bcs/restore?repo_id=${encodeURIComponent(this._restoreRepoId)}&backup_id=${encodeURIComponent(bid)}`, {});
if (!resp?.ok) {
this._restoreError = this._safeText(resp?.message) || "Restore failed.";
} else {
this._restartRequired = !!resp.restart_required;
this._lastActionMsg = "Restore finished. Restart required.";
this._closeRestore();
}
} catch (e) {
this._restoreError = e?.message ? String(e.message) : String(e);
} finally {
this._restoring = false;
await this._load();
}
}
async _restartHA() { async _restartHA() {
if (!this._hass) return; if (!this._hass) return;
try { try {
@@ -273,8 +383,60 @@ class BahmcloudStorePanel extends HTMLElement {
this._readmeExpanded = false; this._readmeExpanded = false;
this._readmeCanToggle = false; this._readmeCanToggle = false;
// Versions dropdown
if (!(repoId in this._selectedVersionByRepoId)) {
this._selectedVersionByRepoId[repoId] = ""; // default = latest
}
this._update(); this._update();
this._loadRepoDetails(repoId);
this._loadReadme(repoId); this._loadReadme(repoId);
this._loadVersions(repoId);
}
async _loadRepoDetails(repoId) {
if (!this._hass || !repoId) return;
try {
const resp = await this._hass.callApi("get", `bcs/repo?repo_id=${encodeURIComponent(repoId)}`);
if (resp?.ok && resp.repo) {
this._detailRepo = resp.repo;
// Also update the cached list item if present
const repos = Array.isArray(this._data?.repos) ? this._data.repos : [];
const idx = repos.findIndex((r) => this._safeId(r?.id) === repoId);
if (idx >= 0) repos[idx] = resp.repo;
this._update();
}
} catch (e) {
// ignore: details are optional
}
}
async _loadVersions(repoId) {
if (!this._hass) return;
if (!repoId) return;
// Cache: avoid re-fetching repeatedly in the same session.
if (Array.isArray(this._versionsCache?.[repoId]) && this._versionsCache[repoId].length) {
return;
}
this._versionsLoadingRepoId = repoId;
this._update();
try {
const resp = await this._hass.callApi("get", `bcs/versions?repo_id=${encodeURIComponent(repoId)}`);
if (resp?.ok && Array.isArray(resp.versions)) {
this._versionsCache[repoId] = resp.versions;
} else {
this._versionsCache[repoId] = [];
}
} catch (e) {
this._versionsCache[repoId] = [];
} finally {
this._versionsLoadingRepoId = null;
this._update();
}
} }
async _loadReadme(repoId) { async _loadReadme(repoId) {
@@ -413,6 +575,22 @@ class BahmcloudStorePanel extends HTMLElement {
} }
button:disabled{ opacity: .55; cursor: not-allowed; } button:disabled{ opacity: .55; cursor: not-allowed; }
.modalOverlay{
position:fixed; inset:0; z-index:999;
background: rgba(0,0,0,0.45);
display:flex; align-items:center; justify-content:center;
padding:16px;
}
.modal{
width: min(520px, 100%);
background: var(--card-background-color);
border:1px solid var(--divider-color);
border-radius:18px;
padding:16px;
box-shadow: 0 10px 30px rgba(0,0,0,0.25);
}
.modalTitle{ font-size:16px; font-weight:700; }
.err{ .err{
margin:12px 0; margin:12px 0;
padding:12px 14px; padding:12px 14px;
@@ -617,7 +795,8 @@ class BahmcloudStorePanel extends HTMLElement {
else if (this._view === "about") html = this._renderAbout(); else if (this._view === "about") html = this._renderAbout();
else if (this._view === "detail") html = this._renderDetail(); else if (this._view === "detail") html = this._renderDetail();
content.innerHTML = `${err}${html}`; const modal = this._renderRestoreModal();
content.innerHTML = `${err}${html}${modal}`;
if (this._view === "store") this._wireStore(); if (this._view === "store") this._wireStore();
if (this._view === "manage") this._wireManage(); if (this._view === "manage") this._wireManage();
@@ -625,6 +804,8 @@ class BahmcloudStorePanel extends HTMLElement {
this._wireDetail(); // now always wires buttons this._wireDetail(); // now always wires buttons
} }
this._wireRestoreModal();
// Restore focus and cursor for the search field after re-render. // Restore focus and cursor for the search field after re-render.
if (restore.id && this._view === "store") { if (restore.id && this._view === "store") {
const el = root.getElementById(restore.id); const el = root.getElementById(restore.id);
@@ -841,6 +1022,8 @@ class BahmcloudStorePanel extends HTMLElement {
const r = this._detailRepo; const r = this._detailRepo;
if (!r) return `<div class="card">No repository selected.</div>`; if (!r) return `<div class="card">No repository selected.</div>`;
const repoId = this._safeId(r?.id) || this._detailRepoId || "";
const name = this._safeText(r?.name) || "Unnamed repository"; const name = this._safeText(r?.name) || "Unnamed repository";
const url = this._safeText(r?.url) || ""; const url = this._safeText(r?.url) || "";
const desc = this._safeText(r?.description) || ""; const desc = this._safeText(r?.description) || "";
@@ -889,8 +1072,6 @@ class BahmcloudStorePanel extends HTMLElement {
</div> </div>
`; `;
const repoId = this._safeId(r?.id);
const installed = this._asBoolStrict(r?.installed); const installed = this._asBoolStrict(r?.installed);
const installedVersion = this._safeText(r?.installed_version); const installedVersion = this._safeText(r?.installed_version);
const installedDomains = Array.isArray(r?.installed_domains) ? r.installed_domains : []; const installedDomains = Array.isArray(r?.installed_domains) ? r.installed_domains : [];
@@ -903,9 +1084,36 @@ class BahmcloudStorePanel extends HTMLElement {
const updateAvailable = installed && !!latestVersion && (!installedVersion || latestVersion !== installedVersion); const updateAvailable = installed && !!latestVersion && (!installedVersion || latestVersion !== installedVersion);
const versions = Array.isArray(this._versionsCache?.[repoId]) ? this._versionsCache[repoId] : [];
const versionsLoading = this._versionsLoadingRepoId === repoId;
const selectedRef = this._safeText(this._selectedVersionByRepoId?.[repoId] || "").trim();
let versionOptions = `<option value="">Latest (recommended)</option>`;
if (selectedRef && !versions.some((v) => this._safeText(v?.ref) === selectedRef)) {
versionOptions += `<option value="${this._esc(selectedRef)}" selected>Selected: ${this._esc(selectedRef)}</option>`;
}
for (const v of versions) {
const ref = this._safeText(v?.ref);
if (!ref) continue;
const label = this._safeText(v?.label) || ref;
const sel = selectedRef === ref ? "selected" : "";
versionOptions += `<option value="${this._esc(ref)}" ${sel}>${this._esc(label)}</option>`;
}
const versionSelect = `
<div style="margin-top:12px;">
<div class="muted small" style="margin-bottom:6px;"><strong>Install version:</strong></div>
<select id="selVersion" ${busy ? "disabled" : ""} style="width:100%;">
${versionOptions}
</select>
${versionsLoading ? `<div class="muted small" style="margin-top:6px;">Loading versions…</div>` : ``}
</div>
`;
const installBtn = `<button class="primary" id="btnInstall" ${installed || busy ? "disabled" : ""}>${busyInstall ? "Installing…" : installed ? "Installed" : "Install"}</button>`; const installBtn = `<button class="primary" id="btnInstall" ${installed || busy ? "disabled" : ""}>${busyInstall ? "Installing…" : installed ? "Installed" : "Install"}</button>`;
const updateBtn = `<button class="primary" id="btnUpdate" ${!updateAvailable || busy ? "disabled" : ""}>${busyUpdate ? "Updating…" : updateAvailable ? "Update" : "Up to date"}</button>`; const updateBtn = `<button class="primary" id="btnUpdate" ${!updateAvailable || busy ? "disabled" : ""}>${busyUpdate ? "Updating…" : updateAvailable ? "Update" : "Up to date"}</button>`;
const uninstallBtn = `<button class="primary" id="btnUninstall" ${!installed || busy ? "disabled" : ""}>${busyUninstall ? "Uninstalling…" : "Uninstall"}</button>`; const uninstallBtn = `<button class="primary" id="btnUninstall" ${!installed || busy ? "disabled" : ""}>${busyUninstall ? "Uninstalling…" : "Uninstall"}</button>`;
const restoreBtn = `<button class="primary" id="btnRestore" ${!installed || busy ? "disabled" : ""}>Restore</button>`;
const restartHint = this._restartRequired const restartHint = this._restartRequired
? ` ? `
@@ -954,10 +1162,13 @@ class BahmcloudStorePanel extends HTMLElement {
<div style="margin-top:6px;"><strong>Domains:</strong> ${installedDomains.length ? this._esc(installedDomains.join(", ")) : "-"}</div> <div style="margin-top:6px;"><strong>Domains:</strong> ${installedDomains.length ? this._esc(installedDomains.join(", ")) : "-"}</div>
</div> </div>
${versionSelect}
<div class="row" style="margin-top:14px; gap:10px; flex-wrap:wrap;"> <div class="row" style="margin-top:14px; gap:10px; flex-wrap:wrap;">
${installBtn} ${installBtn}
${updateBtn} ${updateBtn}
${uninstallBtn} ${uninstallBtn}
${restoreBtn}
</div> </div>
${restartHint} ${restartHint}
@@ -974,8 +1185,10 @@ class BahmcloudStorePanel extends HTMLElement {
const btnInstall = root.getElementById("btnInstall"); const btnInstall = root.getElementById("btnInstall");
const btnUpdate = root.getElementById("btnUpdate"); const btnUpdate = root.getElementById("btnUpdate");
const btnUninstall = root.getElementById("btnUninstall"); const btnUninstall = root.getElementById("btnUninstall");
const btnRestore = root.getElementById("btnRestore");
const btnRestart = root.getElementById("btnRestart"); const btnRestart = root.getElementById("btnRestart");
const btnReadmeToggle = root.getElementById("btnReadmeToggle"); const btnReadmeToggle = root.getElementById("btnReadmeToggle");
const selVersion = root.getElementById("selVersion");
if (btnInstall) { if (btnInstall) {
btnInstall.addEventListener("click", () => { btnInstall.addEventListener("click", () => {
@@ -984,6 +1197,14 @@ class BahmcloudStorePanel extends HTMLElement {
}); });
} }
if (selVersion) {
selVersion.addEventListener("change", () => {
if (!this._detailRepoId) return;
const v = selVersion.value != null ? String(selVersion.value) : "";
this._selectedVersionByRepoId[this._detailRepoId] = v;
});
}
if (btnUpdate) { if (btnUpdate) {
btnUpdate.addEventListener("click", () => { btnUpdate.addEventListener("click", () => {
if (btnUpdate.disabled) return; if (btnUpdate.disabled) return;
@@ -998,6 +1219,15 @@ class BahmcloudStorePanel extends HTMLElement {
}); });
} }
if (btnRestore) {
btnRestore.addEventListener("click", () => {
if (btnRestore.disabled) return;
if (this._detailRepoId) this._openRestore(this._detailRepoId);
});
}
if (btnRestart) { if (btnRestart) {
btnRestart.addEventListener("click", () => this._restartHA()); btnRestart.addEventListener("click", () => this._restartHA());
} }
@@ -1024,6 +1254,45 @@ class BahmcloudStorePanel extends HTMLElement {
} }
} }
_wireRestoreModal() {
const root = this.shadowRoot;
if (!root) return;
const overlay = root.getElementById("restoreOverlay");
if (!overlay) return;
// Click outside modal closes it
overlay.addEventListener("click", (ev) => {
if (ev.target === overlay) this._closeRestore();
});
const sel = root.getElementById("restoreSelect");
if (sel) {
try {
if (this._restoreSelected) sel.value = String(this._restoreSelected);
} catch (_) {}
sel.addEventListener("change", () => {
this._restoreSelected = String(sel.value || "");
});
}
const btnCancel = root.getElementById("btnRestoreCancel");
if (btnCancel) {
btnCancel.addEventListener("click", () => this._closeRestore());
}
const btnApply = root.getElementById("btnRestoreApply");
if (btnApply) {
btnApply.addEventListener("click", () => {
if (btnApply.disabled) return;
const v = root.getElementById("restoreSelect")?.value;
if (v) this._restoreSelected = String(v);
this._restoreSelectedBackup();
});
}
}
_postprocessRenderedMarkdown(container) { _postprocessRenderedMarkdown(container) {
if (!container) return; if (!container) return;
try { try {
@@ -1036,6 +1305,52 @@ class BahmcloudStorePanel extends HTMLElement {
} }
_renderRestoreModal() {
if (!this._restoreOpen) return "";
const opts = Array.isArray(this._restoreOptions) ? this._restoreOptions : [];
const disabled = this._restoreLoading || this._restoring || !opts.length;
const optionsHtml = opts
.map((o) => {
const id = this._safeText(o?.id) || "";
const label = this._safeText(o?.label) || id;
return `<option value="${this._esc(id)}">${this._esc(label)}</option>`;
})
.join("");
const msg = this._restoreLoading
? "Loading backups…"
: this._restoreError
? this._safeText(this._restoreError)
: opts.length
? "Select a backup to restore. This will overwrite files under /config/custom_components and requires a restart."
: "No backups found.";
return `
<div class="modalOverlay" id="restoreOverlay">
<div class="modal">
<div class="modalTitle">Restore from backup</div>
<div class="muted" style="margin-top:8px;">${this._esc(msg)}</div>
<div style="margin-top:14px;">
<label class="muted small" for="restoreSelect">Backup</label><br/>
<select id="restoreSelect" ${disabled ? "disabled" : ""} style="width:100%; margin-top:6px;">
${optionsHtml}
</select>
</div>
<div class="row" style="margin-top:16px; justify-content:flex-end; gap:10px;">
<button id="btnRestoreCancel">Cancel</button>
<button class="primary" id="btnRestoreApply" ${disabled ? "disabled" : ""}>${this._restoring ? "Restoring…" : "Restore"}</button>
</div>
</div>
</div>
`;
}
_renderManage() { _renderManage() {
const repos = Array.isArray(this._data.repos) ? this._data.repos : []; const repos = Array.isArray(this._data.repos) ? this._data.repos : [];
const custom = repos.filter((r) => r?.source === "custom"); const custom = repos.filter((r) => r?.source === "custom");

View File

@@ -505,3 +505,159 @@ async def fetch_readme_markdown(
continue continue
return None return None
async def fetch_repo_versions(
hass: HomeAssistant,
repo_url: str,
*,
provider: str | None = None,
default_branch: str | None = None,
limit: int = 20,
) -> list[dict[str, str]]:
"""List available versions/refs for a repository.
Returns a list of dicts with keys:
- ref: the ref to install (tag/release/branch)
- label: human-friendly label
- source: release|tag|branch
Notes:
- Uses public endpoints (no tokens) for public repositories.
- We prefer releases first (if available), then tags.
"""
repo_url = (repo_url or "").strip()
if not repo_url:
return []
prov = (provider or "").strip().lower() if provider else ""
if not prov:
prov = detect_provider(repo_url)
owner, repo = _split_owner_repo(repo_url)
if not owner or not repo:
return []
session = async_get_clientsession(hass)
headers = {"User-Agent": UA}
out: list[dict[str, str]] = []
seen: set[str] = set()
def _add(ref: str | None, label: str, source: str) -> None:
r = (ref or "").strip()
if not r or r in seen:
return
seen.add(r)
out.append({"ref": r, "label": label, "source": source})
# Always offer default branch as an explicit option.
if default_branch and str(default_branch).strip():
b = str(default_branch).strip()
_add(b, f"Branch: {b}", "branch")
try:
if prov == "github":
# Releases
gh_headers = {"Accept": "application/vnd.github+json", "User-Agent": UA}
data, _ = await _safe_json(
session,
f"https://api.github.com/repos/{owner}/{repo}/releases?per_page={int(limit)}",
headers=gh_headers,
)
if isinstance(data, list):
for r in data:
if not isinstance(r, dict):
continue
tag = r.get("tag_name")
name = r.get("name")
if tag:
lbl = str(tag)
if isinstance(name, str) and name.strip() and name.strip() != str(tag):
lbl = f"{tag}{name.strip()}"
_add(str(tag), lbl, "release")
# Tags
data, _ = await _safe_json(
session,
f"https://api.github.com/repos/{owner}/{repo}/tags?per_page={int(limit)}",
headers=gh_headers,
)
if isinstance(data, list):
for t in data:
if isinstance(t, dict) and t.get("name"):
_add(str(t["name"]), str(t["name"]), "tag")
return out
if prov == "gitlab":
u = urlparse(repo_url.rstrip("/"))
base = f"{u.scheme}://{u.netloc}"
project = quote_plus(f"{owner}/{repo}")
data, _ = await _safe_json(
session,
f"{base}/api/v4/projects/{project}/releases?per_page={int(limit)}",
headers=headers,
)
if isinstance(data, list):
for r in data:
if not isinstance(r, dict):
continue
tag = r.get("tag_name")
name = r.get("name")
if tag:
lbl = str(tag)
if isinstance(name, str) and name.strip() and name.strip() != str(tag):
lbl = f"{tag}{name.strip()}"
_add(str(tag), lbl, "release")
data, _ = await _safe_json(
session,
f"{base}/api/v4/projects/{project}/repository/tags?per_page={int(limit)}",
headers=headers,
)
if isinstance(data, list):
for t in data:
if isinstance(t, dict) and t.get("name"):
_add(str(t["name"]), str(t["name"]), "tag")
return out
# gitea (incl. Bahmcloud)
u = urlparse(repo_url.rstrip("/"))
base = f"{u.scheme}://{u.netloc}"
data, _ = await _safe_json(
session,
f"{base}/api/v1/repos/{owner}/{repo}/releases?limit={int(limit)}",
headers=headers,
)
if isinstance(data, list):
for r in data:
if not isinstance(r, dict):
continue
tag = r.get("tag_name")
name = r.get("name")
if tag:
lbl = str(tag)
if isinstance(name, str) and name.strip() and name.strip() != str(tag):
lbl = f"{tag}{name.strip()}"
_add(str(tag), lbl, "release")
data, _ = await _safe_json(
session,
f"{base}/api/v1/repos/{owner}/{repo}/tags?limit={int(limit)}",
headers=headers,
)
if isinstance(data, list):
for t in data:
if isinstance(t, dict) and t.get("name"):
_add(str(t["name"]), str(t["name"]), "tag")
return out
except Exception:
_LOGGER.debug("fetch_repo_versions failed for %s", repo_url, exc_info=True)
return out

View File

@@ -292,6 +292,27 @@ class BCSReadmeView(HomeAssistantView):
return web.json_response({"ok": True, "readme": md_str, "html": html}) return web.json_response({"ok": True, "readme": md_str, "html": html})
class BCSVersionsView(HomeAssistantView):
url = "/api/bcs/versions"
name = "api:bcs_versions"
requires_auth = True
def __init__(self, core: Any) -> None:
self.core: BCSCore = core
async def get(self, request: web.Request) -> web.Response:
repo_id = request.query.get("repo_id")
if not repo_id:
return web.json_response({"ok": False, "message": "Missing repo_id"}, status=400)
try:
versions = await self.core.list_repo_versions(repo_id)
return web.json_response({"ok": True, "repo_id": repo_id, "versions": versions}, status=200)
except Exception as e:
_LOGGER.exception("BCS list versions failed: %s", e)
return web.json_response({"ok": False, "message": str(e) or "List versions failed"}, status=500)
class BCSInstallView(HomeAssistantView): class BCSInstallView(HomeAssistantView):
url = "/api/bcs/install" url = "/api/bcs/install"
name = "api:bcs_install" name = "api:bcs_install"
@@ -302,11 +323,13 @@ class BCSInstallView(HomeAssistantView):
async def post(self, request: web.Request) -> web.Response: async def post(self, request: web.Request) -> web.Response:
repo_id = request.query.get("repo_id") repo_id = request.query.get("repo_id")
version = request.query.get("version")
if not repo_id: if not repo_id:
return web.json_response({"ok": False, "message": "Missing repo_id"}, status=400) return web.json_response({"ok": False, "message": "Missing repo_id"}, status=400)
try: try:
result = await self.core.install_repo(repo_id) v = str(version).strip() if version is not None else None
result = await self.core.install_repo(repo_id, version=v)
return web.json_response(result, status=200) return web.json_response(result, status=200)
except Exception as e: except Exception as e:
_LOGGER.exception("BCS install failed: %s", e) _LOGGER.exception("BCS install failed: %s", e)
@@ -323,11 +346,13 @@ class BCSUpdateView(HomeAssistantView):
async def post(self, request: web.Request) -> web.Response: async def post(self, request: web.Request) -> web.Response:
repo_id = request.query.get("repo_id") repo_id = request.query.get("repo_id")
version = request.query.get("version")
if not repo_id: if not repo_id:
return web.json_response({"ok": False, "message": "Missing repo_id"}, status=400) return web.json_response({"ok": False, "message": "Missing repo_id"}, status=400)
try: try:
result = await self.core.update_repo(repo_id) v = str(version).strip() if version is not None else None
result = await self.core.update_repo(repo_id, version=v)
return web.json_response(result, status=200) return web.json_response(result, status=200)
except Exception as e: except Exception as e:
_LOGGER.exception("BCS update failed: %s", e) _LOGGER.exception("BCS update failed: %s", e)
@@ -355,6 +380,53 @@ class BCSUninstallView(HomeAssistantView):
return web.json_response({"ok": False, "message": str(e) or "Uninstall failed"}, status=500) return web.json_response({"ok": False, "message": str(e) or "Uninstall failed"}, status=500)
class BCSBackupsView(HomeAssistantView):
url = "/api/bcs/backups"
name = "api:bcs_backups"
requires_auth = True
def __init__(self, core: Any) -> None:
self.core: BCSCore = core
async def get(self, request: web.Request) -> web.Response:
repo_id = request.query.get("repo_id")
if not repo_id:
return web.json_response({"ok": False, "message": "Missing repo_id"}, status=400)
try:
backups = await self.core.list_repo_backups(repo_id)
return web.json_response({"ok": True, "repo_id": repo_id, "backups": backups}, status=200)
except Exception as e:
_LOGGER.exception("BCS list backups failed: %s", e)
return web.json_response({"ok": False, "message": str(e) or "List backups failed"}, status=500)
class BCSRestoreView(HomeAssistantView):
url = "/api/bcs/restore"
name = "api:bcs_restore"
requires_auth = True
def __init__(self, core: Any) -> None:
self.core: BCSCore = core
async def post(self, request: web.Request) -> web.Response:
repo_id = request.query.get("repo_id")
backup_id = request.query.get("backup_id")
if not repo_id:
return web.json_response({"ok": False, "message": "Missing repo_id"}, status=400)
if not backup_id:
return web.json_response({"ok": False, "message": "Missing backup_id"}, status=400)
try:
result = await self.core.restore_repo_backup(repo_id, backup_id)
return web.json_response(result, status=200)
except Exception as e:
_LOGGER.exception("BCS restore failed: %s", e)
return web.json_response({"ok": False, "message": str(e) or "Restore failed"}, status=500)
class BCSRestartView(HomeAssistantView): class BCSRestartView(HomeAssistantView):
url = "/api/bcs/restart" url = "/api/bcs/restart"
name = "api:bcs_restart" name = "api:bcs_restart"
@@ -370,3 +442,55 @@ class BCSRestartView(HomeAssistantView):
except Exception as e: except Exception as e:
_LOGGER.exception("BCS restart failed: %s", e) _LOGGER.exception("BCS restart failed: %s", e)
return web.json_response({"ok": False, "message": str(e) or "Restart failed"}, status=500) return web.json_response({"ok": False, "message": str(e) or "Restart failed"}, status=500)
class BCSRepoDetailView(HomeAssistantView):
url = "/api/bcs/repo"
name = "api:bcs_repo"
requires_auth = True
def __init__(self, core: Any) -> None:
self.core: BCSCore = core
async def get(self, request: web.Request) -> web.Response:
repo_id = (request.query.get("repo_id") or "").strip()
if not repo_id:
return web.json_response({"ok": False, "message": "Missing repo_id"}, status=400)
try:
repo = await self.core.ensure_repo_details(repo_id)
if not repo:
return web.json_response({"ok": False, "message": "Repo not found"}, status=404)
inst = self.core.get_installed(repo_id) or {}
installed = bool(inst)
domains = inst.get("domains") or []
if not isinstance(domains, list):
domains = []
return web.json_response({
"ok": True,
"repo": {
"id": repo.id,
"name": repo.name,
"url": repo.url,
"source": repo.source,
"owner": repo.owner,
"provider": repo.provider,
"repo_name": repo.provider_repo_name,
"description": repo.provider_description or repo.meta_description,
"default_branch": repo.default_branch,
"latest_version": repo.latest_version,
"latest_version_source": repo.latest_version_source,
"category": repo.meta_category,
"meta_author": repo.meta_author,
"meta_maintainer": repo.meta_maintainer,
"meta_source": repo.meta_source,
"installed": installed,
"installed_version": inst.get("installed_version"),
"installed_manifest_version": inst.get("installed_manifest_version"),
"installed_domains": domains,
}
}, status=200)
except Exception as e:
_LOGGER.exception("BCS repo details failed: %s", e)
return web.json_response({"ok": False, "message": str(e) or "Repo details failed"}, status=500)